+ combrec=1 ++ dirname ./scripts/raidz_expand_test.sh + BASE_DIR=./scripts/.. + echo 1 + zpool destroy test cannot open 'test': no such pool + zpool create filepool sdb + zfs destroy -R filepool/files cannot open 'filepool/files': dataset does not exist + zfs create -o compression=on filepool/files + dir=/filepool/files + (( i=0 )) + (( i<7 )) + truncate -s 512M /filepool/files/0 + (( i=i+1 )) + (( i<7 )) + truncate -s 512M /filepool/files/1 + (( i=i+1 )) + (( i<7 )) + truncate -s 512M /filepool/files/2 + (( i=i+1 )) + (( i<7 )) + truncate -s 512M /filepool/files/3 + (( i=i+1 )) + (( i<7 )) + truncate -s 512M /filepool/files/4 + (( i=i+1 )) + (( i<7 )) + truncate -s 512M /filepool/files/5 + (( i=i+1 )) + (( i<7 )) + truncate -s 512M /filepool/files/6 + (( i=i+1 )) + (( i<7 )) + dotest 2 + nparity=2 + zpool create -o cachefile=none test raidz2 /filepool/files/0 /filepool/files/1 /filepool/files/2 /filepool/files/3 /filepool/files/4 /filepool/files/5 + zfs set primarycache=metadata test + zfs create test/fs + dd if=/dev/urandom of=/test/fs/file bs=1024k count=1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00613063 s, 171 MB/s + zfs create -o compress=on test/fs2 + cp -r ./scripts/.. /test/fs2/ + zfs create -o compress=on -o recordsize=8k test/fs3 + cp -r ./scripts/.. /test/fs3/ + zfs snapshot filepool/files@pre-attach + sum /test/fs/file 21315 1024 + sum /test/fs2/file sum: /test/fs2/file: No such file or directory + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zfs list test NAME USED AVAIL REFER MOUNTPOINT test 400M 1.32G 46.0K /test + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 2.75G 601M 2.16G - - 0% 21% 1.00x ONLINE - raidz2 2.75G 601M 2.16G - - 0% 21.4% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE + sleep 2 + zpool attach test raidz2-0 /filepool/files/6 ***************************************************** * Thank you for testing this alpha-quality release * * of RAID-Z expansion. This feature should only * * be used on test pools. The pool will eventually * * need to be DESTROYED, because the on-disk format * * will not be compatible with the final release. * * Additionally, there are currently bugs in RAID-Z * * expansion which can occasionally cause data loss. * * Please report bugs to mahrens@delphix.com. * ***************************************************** You have 5 seconds to abort by pressing ^C (control-C) You have 4 seconds to abort by pressing ^C (control-C) You have 3 seconds to abort by pressing ^C (control-C) You have 2 seconds to abort by pressing ^C (control-C) You have 1 seconds to abort by pressing ^C (control-C) + wait_completion + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:21:46 2020 + sleep 5 + grep 'in progress' + zpool status test + zfs list test NAME USED AVAIL REFER MOUNTPOINT test 404M 1.64G 46.0K /test + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 3.25G 607M 2.66G - - 0% 18% 1.00x ONLINE - raidz2 3.25G 607M 2.66G - - 0% 18.2% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE /filepool/files/6 - - - - - - - - ONLINE + zfs snapshot filepool/files@post-attach + zpool export test + zpool import -o cachefile=none -d /filepool/files test + zfs snapshot filepool/files@post-import + sum /test/fs/file 21315 1024 + sum /test/fs2/file sum: /test/fs2/file: No such file or directory + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zfs list -r test NAME USED AVAIL REFER MOUNTPOINT test 404M 1.65G 46.0K /test test/fs 1.04M 1.65G 1.04M /test/fs test/fs2 182M 1.65G 182M /test/fs2 test/fs3 221M 1.65G 221M /test/fs3 + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 3.25G 606M 2.66G - - 0% 18% 1.00x ONLINE - raidz2 3.25G 606M 2.66G - - 0% 18.2% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE /filepool/files/6 - - - - - - - - ONLINE + zpool status -v test pool: test state: ONLINE scan: none requested raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:22:30 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE scan: scrub repaired 0B in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:22:34 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool export test + zpool import -o cachefile=none -d /filepool/files test + (( i=0 )) + (( i<2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/0 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.211584 s, 2.5 GB/s + (( i=i+1 )) + (( i<2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/1 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.173921 s, 3.0 GB/s + (( i=i+1 )) + (( i<2 )) + sum /test/fs/file 21315 1024 + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 0B in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:22:34 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 440 /filepool/files/1 ONLINE 0 0 438 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + [[ -n 1 ]] + zpool scrub test + wait_completion + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:22:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:22:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:22:38 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 166M in 0 days 00:00:11 with 0 errors on Fri Apr 3 20:22:49 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 166K /filepool/files/1 ONLINE 0 0 166K /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool clear test + (( i=2 )) + (( i<2*2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/2 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.15896 s, 3.3 GB/s + (( i=i+1 )) + (( i<2*2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/3 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.165465 s, 3.2 GB/s + (( i=i+1 )) + (( i<2*2 )) + zpool status -v test pool: test state: ONLINE scan: scrub repaired 166M in 0 days 00:00:11 with 0 errors on Fri Apr 3 20:22:49 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + [[ -n 1 ]] + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:22:54 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:22:54 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:22:54 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 166M in 0 days 00:00:14 with 0 errors on Fri Apr 3 20:23:08 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 166K /filepool/files/3 ONLINE 0 0 166K /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool clear test + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:23:10 2020 + sleep 5 + grep 'in progress' + zpool status test + zpool status -v test pool: test state: ONLINE scan: scrub repaired 0B in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:23:14 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:22:26 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool destroy test + dotest 3 + nparity=3 + zpool create -o cachefile=none test raidz3 /filepool/files/0 /filepool/files/1 /filepool/files/2 /filepool/files/3 /filepool/files/4 /filepool/files/5 + zfs set primarycache=metadata test + zfs create test/fs + dd if=/dev/urandom of=/test/fs/file bs=1024k count=1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00637303 s, 165 MB/s + zfs create -o compress=on test/fs2 + cp -r ./scripts/.. /test/fs2/ + zfs create -o compress=on -o recordsize=8k test/fs3 + cp -r ./scripts/.. /test/fs3/ + zfs snapshot filepool/files@pre-attach cannot create snapshot 'filepool/files@pre-attach': dataset already exists + sum /test/fs/file 64111 1024 + sum /test/fs2/file sum: /test/fs2/file: No such file or directory + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zfs list test NAME USED AVAIL REFER MOUNTPOINT test 408M 861M 45.6K /test + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 2.75G 823M 1.95G - - 0% 29% 1.00x ONLINE - raidz3 2.75G 823M 1.95G - - 0% 29.2% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE + sleep 2 + zpool attach test raidz3-0 /filepool/files/6 ***************************************************** * Thank you for testing this alpha-quality release * * of RAID-Z expansion. This feature should only * * be used on test pools. The pool will eventually * * need to be DESTROYED, because the on-disk format * * will not be compatible with the final release. * * Additionally, there are currently bugs in RAID-Z * * expansion which can occasionally cause data loss. * * Please report bugs to mahrens@delphix.com. * ***************************************************** You have 5 seconds to abort by pressing ^C (control-C) You have 4 seconds to abort by pressing ^C (control-C) You have 3 seconds to abort by pressing ^C (control-C) You have 2 seconds to abort by pressing ^C (control-C) You have 1 seconds to abort by pressing ^C (control-C) + wait_completion + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:23:41 2020 + sleep 5 + grep 'in progress' + zpool status test + zfs list test NAME USED AVAIL REFER MOUNTPOINT test 411M 1.09G 45.6K /test + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 3.25G 828M 2.44G - - 0% 24% 1.00x ONLINE - raidz3 3.25G 828M 2.44G - - 0% 24.9% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE /filepool/files/6 - - - - - - - - ONLINE + zfs snapshot filepool/files@post-attach cannot create snapshot 'filepool/files@post-attach': dataset already exists + zpool export test + zpool import -o cachefile=none -d /filepool/files test + zfs snapshot filepool/files@post-import cannot create snapshot 'filepool/files@post-import': dataset already exists + sum /test/fs/file 64111 1024 + sum /test/fs2/file sum: /test/fs2/file: No such file or directory + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zfs list -r test NAME USED AVAIL REFER MOUNTPOINT test 410M 1.09G 45.6K /test test/fs 1.04M 1.09G 1.04M /test/fs test/fs2 182M 1.09G 182M /test/fs2 test/fs3 227M 1.09G 227M /test/fs3 + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 3.25G 827M 2.44G - - 0% 24% 1.00x ONLINE - raidz3 3.25G 827M 2.44G - - 0% 24.8% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE /filepool/files/6 - - - - - - - - ONLINE + zpool status -v test pool: test state: ONLINE scan: none requested raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool scrub test + wait_completion + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:24:30 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE scan: scrub repaired 0B in 0 days 00:00:05 with 0 errors on Fri Apr 3 20:24:35 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool export test + zpool import -o cachefile=none -d /filepool/files test + (( i=0 )) + (( i<3 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/0 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.204684 s, 2.6 GB/s + (( i=i+1 )) + (( i<3 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/1 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.16421 s, 3.2 GB/s + (( i=i+1 )) + (( i<3 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/2 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.170031 s, 3.1 GB/s + (( i=i+1 )) + (( i<3 )) + sum /test/fs/file 64111 1024 + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 0B in 0 days 00:00:05 with 0 errors on Fri Apr 3 20:24:35 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 589 /filepool/files/1 ONLINE 0 0 588 /filepool/files/2 ONLINE 0 0 589 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + [[ -n 1 ]] + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:24:38 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 334M in 0 days 00:01:04 with 0 errors on Fri Apr 3 20:25:42 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 224K /filepool/files/1 ONLINE 0 0 223K /filepool/files/2 ONLINE 0 0 224K /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 10 errors: No known data errors + zpool clear test + (( i=3 )) + (( i<3*2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/3 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.199114 s, 2.6 GB/s + (( i=i+1 )) + (( i<3*2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/4 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.15346 s, 3.4 GB/s + (( i=i+1 )) + (( i<3*2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/5 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.198215 s, 2.6 GB/s + (( i=i+1 )) + (( i<3*2 )) + zpool status -v test pool: test state: ONLINE scan: scrub repaired 334M in 0 days 00:01:04 with 0 errors on Fri Apr 3 20:25:42 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + [[ -n 1 ]] + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test scan: scrub in progress since Fri Apr 3 20:25:45 2020 + sleep 5 + grep 'in progress' + zpool status test + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 335M in 0 days 00:01:22 with 0 errors on Fri Apr 3 20:27:07 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 8 /filepool/files/3 ONLINE 0 0 224K /filepool/files/4 ONLINE 0 0 224K /filepool/files/5 ONLINE 0 0 224K /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool clear test + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:27:08 2020 + sleep 5 + grep 'in progress' + zpool status test + zpool status -v test pool: test state: ONLINE scan: scrub repaired 0B in 0 days 00:00:05 with 0 errors on Fri Apr 3 20:27:13 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:24:23 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool destroy test + dotest 1 + nparity=1 + zpool create -o cachefile=none test raidz1 /filepool/files/0 /filepool/files/1 /filepool/files/2 /filepool/files/3 /filepool/files/4 /filepool/files/5 + zfs set primarycache=metadata test + zfs create test/fs + dd if=/dev/urandom of=/test/fs/file bs=1024k count=1 1+0 records in 1+0 records out 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00390965 s, 268 MB/s + zfs create -o compress=on test/fs2 + cp -r ./scripts/.. /test/fs2/ + zfs create -o compress=on -o recordsize=8k test/fs3 + cp -r ./scripts/.. /test/fs3/ + zfs snapshot filepool/files@pre-attach cannot create snapshot 'filepool/files@pre-attach': dataset already exists + sum /test/fs/file 51037 1024 + sum /test/fs2/file sum: /test/fs2/file: No such file or directory + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zfs list test NAME USED AVAIL REFER MOUNTPOINT test 388M 1.78G 39.8K /test + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 2.75G 468M 2.29G - - 0% 16% 1.00x ONLINE - raidz1 2.75G 468M 2.29G - - 0% 16.6% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE + sleep 2 + zpool attach test raidz1-0 /filepool/files/6 ***************************************************** * Thank you for testing this alpha-quality release * * of RAID-Z expansion. This feature should only * * be used on test pools. The pool will eventually * * need to be DESTROYED, because the on-disk format * * will not be compatible with the final release. * * Additionally, there are currently bugs in RAID-Z * * expansion which can occasionally cause data loss. * * Please report bugs to mahrens@delphix.com. * ***************************************************** You have 5 seconds to abort by pressing ^C (control-C) You have 4 seconds to abort by pressing ^C (control-C) You have 3 seconds to abort by pressing ^C (control-C) You have 2 seconds to abort by pressing ^C (control-C) You have 1 seconds to abort by pressing ^C (control-C) + wait_completion + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + grep 'in progress' + zpool status test raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + zpool status test + grep 'in progress' raidz expand: Expansion of vdev 0 in progress since Fri Apr 3 20:27:33 2020 + sleep 5 + zpool status test + grep 'in progress' + zfs list test NAME USED AVAIL REFER MOUNTPOINT test 395M 2.19G 39.8K /test + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 3.25G 476M 2.78G - - 0% 14% 1.00x ONLINE - raidz1 3.25G 476M 2.78G - - 0% 14.3% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE /filepool/files/6 - - - - - - - - ONLINE + zfs snapshot filepool/files@post-attach cannot create snapshot 'filepool/files@post-attach': dataset already exists + zpool export test + zpool import -o cachefile=none -d /filepool/files test + zfs snapshot filepool/files@post-import cannot create snapshot 'filepool/files@post-import': dataset already exists + sum /test/fs/file 51037 1024 + sum /test/fs2/file sum: /test/fs2/file: No such file or directory + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zfs list -r test NAME USED AVAIL REFER MOUNTPOINT test 395M 2.19G 39.8K /test test/fs 1.04M 2.19G 1.04M /test/fs test/fs2 179M 2.19G 179M /test/fs2 test/fs3 214M 2.19G 214M /test/fs3 + zpool list -v test NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT test 3.25G 476M 2.79G - - 0% 14% 1.00x ONLINE - raidz1 3.25G 476M 2.79G - - 0% 14.3% - ONLINE /filepool/files/0 - - - - - - - - ONLINE /filepool/files/1 - - - - - - - - ONLINE /filepool/files/2 - - - - - - - - ONLINE /filepool/files/3 - - - - - - - - ONLINE /filepool/files/4 - - - - - - - - ONLINE /filepool/files/5 - - - - - - - - ONLINE /filepool/files/6 - - - - - - - - ONLINE + zpool status -v test pool: test state: ONLINE scan: none requested raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:28:16 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE scan: scrub repaired 0B in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:28:20 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool export test + zpool import -o cachefile=none -d /filepool/files test + (( i=0 )) + (( i<1 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/0 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.182371 s, 2.9 GB/s + (( i=i+1 )) + (( i<1 )) + sum /test/fs/file 51037 1024 + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 0B in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:28:20 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 352 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + [[ -n 1 ]] + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:28:24 2020 + sleep 5 + grep 'in progress' + zpool status test + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 66.1M in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:28:28 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 133K /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool clear test + (( i=1 )) + (( i<1*2 )) + [[ ! -n 1 ]] + dd conv=notrunc if=/dev/zero of=/filepool/files/1 bs=1024k seek=4 count=500 500+0 records in 500+0 records out 524288000 bytes (524 MB, 500 MiB) copied, 0.181322 s, 2.9 GB/s + (( i=i+1 )) + (( i<1*2 )) + zpool status -v test pool: test state: ONLINE scan: scrub repaired 66.1M in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:28:28 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + [[ -n 1 ]] + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:28:30 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: https://zfsonlinux.org/msg/ZFS-8000-9P scan: scrub repaired 66.3M in 0 days 00:00:04 with 0 errors on Fri Apr 3 20:28:34 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 133K /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool clear test + sum /test/fs3/file sum: /test/fs3/file: No such file or directory + zpool scrub test + wait_completion + zpool status test + grep 'in progress' scan: scrub in progress since Fri Apr 3 20:28:36 2020 + sleep 5 + zpool status test + grep 'in progress' + zpool status -v test pool: test state: ONLINE scan: scrub repaired 0B in 0 days 00:00:03 with 0 errors on Fri Apr 3 20:28:39 2020 raidz expand: Expansion of vdev 0 copied 3.47G in 0h0m, completed on Fri Apr 3 20:28:10 2020 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 /filepool/files/0 ONLINE 0 0 0 /filepool/files/1 ONLINE 0 0 0 /filepool/files/2 ONLINE 0 0 0 /filepool/files/3 ONLINE 0 0 0 /filepool/files/4 ONLINE 0 0 0 /filepool/files/5 ONLINE 0 0 0 /filepool/files/6 ONLINE 0 0 0 errors: No known data errors + zpool destroy test