Message ID | 20180620084114.25648-2-zlang@redhat.com (mailing list archive) |
---|---|
State | Superseded, archived |
Headers | show |
On Wed, Jun 20, 2018 at 04:41:13PM +0800, Zorro Lang wrote: > This case does dedupe on a dir, then copy the dir to next dir. Dedupe > the next dir again, then copy this dir to next again, and dedupe > again ... At the end, verify the data in the last dir is still same > with the first one. > > Signed-off-by: Zorro Lang <zlang@redhat.com> Same SPDX comment as before but otherwise looks ok, Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com> --D > --- > > V2 did below changes: > 1) Added more description at the case beginning > 2) Changed $TEST_DIR/${seq}md5.sum to $tmp.md5sum > 3) Changed for ...;do format > 4) Remove "-f mknod=0" fsstress option > 5) Added some noise (by fsstress) in each test round. > > Thanks, > Zorro > > tests/shared/009 | 119 +++++++++++++++++++++++++++++++++++++++++++++++++++ > tests/shared/009.out | 4 ++ > tests/shared/group | 1 + > 3 files changed, 124 insertions(+) > create mode 100755 tests/shared/009 > create mode 100644 tests/shared/009.out > > diff --git a/tests/shared/009 b/tests/shared/009 > new file mode 100755 > index 00000000..5ed9faee > --- /dev/null > +++ b/tests/shared/009 > @@ -0,0 +1,119 @@ > +#! /bin/bash > +# FS QA Test 009 > +# > +# Iterate dedupe integrity test. Copy an original data0 several > +# times (d0 -> d1, d1 -> d2, ... dn-1 -> dn), dedupe dataN everytime > +# before copy. At last, verify dataN same with data0. > +# > +#----------------------------------------------------------------------- > +# Copyright (c) 2018 Red Hat Inc. All Rights Reserved. > +# > +# This program is free software; you can redistribute it and/or > +# modify it under the terms of the GNU General Public License as > +# published by the Free Software Foundation. > +# > +# This program is distributed in the hope that it would be useful, > +# but WITHOUT ANY WARRANTY; without even the implied warranty of > +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > +# GNU General Public License for more details. > +# > +# You should have received a copy of the GNU General Public License > +# along with this program; if not, write the Free Software Foundation, > +# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA > +#----------------------------------------------------------------------- > +# > + > +seq=`basename $0` > +seqres=$RESULT_DIR/$seq > +echo "QA output created by $seq" > + > +here=`pwd` > +tmp=/tmp/$$ > +status=1 # failure is the default! > +trap "_cleanup; exit \$status" 0 1 2 3 15 > + > +_cleanup() > +{ > + cd / > + rm -f $tmp.* > +} > + > +# get standard environment, filters and checks > +. ./common/rc > +. ./common/filter > +. ./common/reflink > + > +# remove previous $seqres.full before test > +rm -f $seqres.full > + > +# real QA test starts here > + > +# duperemove only supports btrfs and xfs (with reflink feature). > +# Add other filesystems if it supports more later. > +_supported_fs xfs btrfs > +_supported_os Linux > +_require_scratch_dedupe > +_require_command "$DUPEREMOVE_PROG" duperemove > + > +_scratch_mkfs > $seqres.full 2>&1 > +_scratch_mount >> $seqres.full 2>&1 > + > +function iterate_dedup_verify() > +{ > + local src=$srcdir > + local dest=$dupdir/1 > + > + for ((index = 1; index <= times; index++)); do > + cp -a $src $dest > + find $dest -type f -exec md5sum {} \; \ > + > $md5file$index > + # Make some noise > + $FSSTRESS_PROG $fsstress_opts -d $noisedir \ > + -n 200 -p $((5 * LOAD_FACTOR)) >/dev/null 2>&1 > + # Too many output, so only save error output > + $DUPEREMOVE_PROG -dr --dedupe-options=same $dupdir \ > + >/dev/null 2>$seqres.full > + md5sum -c --quiet $md5file$index > + src=$dest > + dest=$dupdir/$((index + 1)) > + done > +} > + > +srcdir=$SCRATCH_MNT/src > +dupdir=$SCRATCH_MNT/dup > +noisedir=$dupdir/noise > +mkdir $srcdir $dupdir > +mkdir $dupdir/noise > + > +md5file=${tmp}.md5sum > + > +fsstress_opts="-w -r" > +# Create some files to be original data > +$FSSTRESS_PROG $fsstress_opts -d $srcdir \ > + -n 500 -p $((5 * LOAD_FACTOR)) >/dev/null 2>&1 > + > +# Calculate how many test cycles will be run > +src_size=`du -ks $srcdir | awk '{print $1}'` > +free_size=`df -kP $SCRATCH_MNT | grep -v Filesystem | awk '{print $4}'` > +times=$((free_size / src_size)) > +if [ $times -gt $((4 * TIME_FACTOR)) ]; then > + times=$((4 * TIME_FACTOR)) > +fi > + > +echo "= Do dedup and verify =" > +iterate_dedup_verify > + > +# Use the last checksum file to verify the original data > +sed -e s#dup/$times#src#g $md5file$times > $md5file > +echo "= Backwords verify =" > +md5sum -c --quiet $md5file > + > +# read from the disk also doesn't show mutations. > +_scratch_cycle_mount > +echo "= Verify after cycle mount =" > +for ((index = 1; index <= times; index++)); do > + md5sum -c --quiet $md5file$index > +done > + > +status=0 > +exit > diff --git a/tests/shared/009.out b/tests/shared/009.out > new file mode 100644 > index 00000000..44a78ba3 > --- /dev/null > +++ b/tests/shared/009.out > @@ -0,0 +1,4 @@ > +QA output created by 009 > += Do dedup and verify = > += Backwords verify = > += Verify after cycle mount = > diff --git a/tests/shared/group b/tests/shared/group > index 49ffa8dd..9c484794 100644 > --- a/tests/shared/group > +++ b/tests/shared/group > @@ -11,6 +11,7 @@ > 006 auto enospc > 007 dangerous_fuzzers > 008 auto stress dedupe > +009 auto stress dedupe > 032 mkfs auto quick > 272 auto enospc rw > 289 auto quick > -- > 2.14.4 > > -- > To unsubscribe from this list: send the line "unsubscribe fstests" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/tests/shared/009 b/tests/shared/009 new file mode 100755 index 00000000..5ed9faee --- /dev/null +++ b/tests/shared/009 @@ -0,0 +1,119 @@ +#! /bin/bash +# FS QA Test 009 +# +# Iterate dedupe integrity test. Copy an original data0 several +# times (d0 -> d1, d1 -> d2, ... dn-1 -> dn), dedupe dataN everytime +# before copy. At last, verify dataN same with data0. +# +#----------------------------------------------------------------------- +# Copyright (c) 2018 Red Hat Inc. All Rights Reserved. +# +# This program is free software; you can redistribute it and/or +# modify it under the terms of the GNU General Public License as +# published by the Free Software Foundation. +# +# This program is distributed in the hope that it would be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program; if not, write the Free Software Foundation, +# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA +#----------------------------------------------------------------------- +# + +seq=`basename $0` +seqres=$RESULT_DIR/$seq +echo "QA output created by $seq" + +here=`pwd` +tmp=/tmp/$$ +status=1 # failure is the default! +trap "_cleanup; exit \$status" 0 1 2 3 15 + +_cleanup() +{ + cd / + rm -f $tmp.* +} + +# get standard environment, filters and checks +. ./common/rc +. ./common/filter +. ./common/reflink + +# remove previous $seqres.full before test +rm -f $seqres.full + +# real QA test starts here + +# duperemove only supports btrfs and xfs (with reflink feature). +# Add other filesystems if it supports more later. +_supported_fs xfs btrfs +_supported_os Linux +_require_scratch_dedupe +_require_command "$DUPEREMOVE_PROG" duperemove + +_scratch_mkfs > $seqres.full 2>&1 +_scratch_mount >> $seqres.full 2>&1 + +function iterate_dedup_verify() +{ + local src=$srcdir + local dest=$dupdir/1 + + for ((index = 1; index <= times; index++)); do + cp -a $src $dest + find $dest -type f -exec md5sum {} \; \ + > $md5file$index + # Make some noise + $FSSTRESS_PROG $fsstress_opts -d $noisedir \ + -n 200 -p $((5 * LOAD_FACTOR)) >/dev/null 2>&1 + # Too many output, so only save error output + $DUPEREMOVE_PROG -dr --dedupe-options=same $dupdir \ + >/dev/null 2>$seqres.full + md5sum -c --quiet $md5file$index + src=$dest + dest=$dupdir/$((index + 1)) + done +} + +srcdir=$SCRATCH_MNT/src +dupdir=$SCRATCH_MNT/dup +noisedir=$dupdir/noise +mkdir $srcdir $dupdir +mkdir $dupdir/noise + +md5file=${tmp}.md5sum + +fsstress_opts="-w -r" +# Create some files to be original data +$FSSTRESS_PROG $fsstress_opts -d $srcdir \ + -n 500 -p $((5 * LOAD_FACTOR)) >/dev/null 2>&1 + +# Calculate how many test cycles will be run +src_size=`du -ks $srcdir | awk '{print $1}'` +free_size=`df -kP $SCRATCH_MNT | grep -v Filesystem | awk '{print $4}'` +times=$((free_size / src_size)) +if [ $times -gt $((4 * TIME_FACTOR)) ]; then + times=$((4 * TIME_FACTOR)) +fi + +echo "= Do dedup and verify =" +iterate_dedup_verify + +# Use the last checksum file to verify the original data +sed -e s#dup/$times#src#g $md5file$times > $md5file +echo "= Backwords verify =" +md5sum -c --quiet $md5file + +# read from the disk also doesn't show mutations. +_scratch_cycle_mount +echo "= Verify after cycle mount =" +for ((index = 1; index <= times; index++)); do + md5sum -c --quiet $md5file$index +done + +status=0 +exit diff --git a/tests/shared/009.out b/tests/shared/009.out new file mode 100644 index 00000000..44a78ba3 --- /dev/null +++ b/tests/shared/009.out @@ -0,0 +1,4 @@ +QA output created by 009 += Do dedup and verify = += Backwords verify = += Verify after cycle mount = diff --git a/tests/shared/group b/tests/shared/group index 49ffa8dd..9c484794 100644 --- a/tests/shared/group +++ b/tests/shared/group @@ -11,6 +11,7 @@ 006 auto enospc 007 dangerous_fuzzers 008 auto stress dedupe +009 auto stress dedupe 032 mkfs auto quick 272 auto enospc rw 289 auto quick
This case does dedupe on a dir, then copy the dir to next dir. Dedupe the next dir again, then copy this dir to next again, and dedupe again ... At the end, verify the data in the last dir is still same with the first one. Signed-off-by: Zorro Lang <zlang@redhat.com> --- V2 did below changes: 1) Added more description at the case beginning 2) Changed $TEST_DIR/${seq}md5.sum to $tmp.md5sum 3) Changed for ...;do format 4) Remove "-f mknod=0" fsstress option 5) Added some noise (by fsstress) in each test round. Thanks, Zorro tests/shared/009 | 119 +++++++++++++++++++++++++++++++++++++++++++++++++++ tests/shared/009.out | 4 ++ tests/shared/group | 1 + 3 files changed, 124 insertions(+) create mode 100755 tests/shared/009 create mode 100644 tests/shared/009.out