From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-17.5 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_CR_TRAILER,INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 03789C433DB for ; Fri, 12 Mar 2021 16:18:35 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id CB0B965019 for ; Fri, 12 Mar 2021 16:18:34 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231557AbhCLQSB (ORCPT ); Fri, 12 Mar 2021 11:18:01 -0500 Received: from us-smtp-delivery-124.mimecast.com ([63.128.21.124]:21323 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231331AbhCLQR6 (ORCPT ); Fri, 12 Mar 2021 11:17:58 -0500 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1615565878; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=W3Xl6WGh2cSXl1txnEKr1O0h3w5ovNgGhu8gztT//y8=; b=Ji0SqmT2vEJZObWlizwMu3kACLRrT+ftrDujL8vDY68ENwzm1NP8AuY3JAB9icuPCxuBhf hlsrw3HRxruXt/7bqqwqXY04p45GkK5spQA0N3WzY2otqoeIqjTmRNnyDCjADWcyO/wiDj A3ZDyy9EUOa2TyPiZpPojfPI4U27RCI= Received: from mail-pl1-f200.google.com (mail-pl1-f200.google.com [209.85.214.200]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-250-mG_qm0vTOvmslzKrxuu7bw-1; Fri, 12 Mar 2021 11:17:56 -0500 X-MC-Unique: mG_qm0vTOvmslzKrxuu7bw-1 Received: by mail-pl1-f200.google.com with SMTP id f13so13522128plj.5 for ; Fri, 12 Mar 2021 08:17:56 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=W3Xl6WGh2cSXl1txnEKr1O0h3w5ovNgGhu8gztT//y8=; b=n1WHgMlIfRXwxrvdl28wnvZMaSy9SjH2cCTDHN/3CQ6QYNOIxhhnxeG+fic/L+XxTB AC6PM6EhTVbExNBML0/yTZUPHMo0Lzt5TksP42Um9lasa4kiJ7QTDU0OeHTfrHmSeRfa BF/3QccKAP/WE7Lxk9dKxRJsfv5wk7hfGa419LApkmS8bAviQZzVgIFbh/ymVfx3AX6I wtmnVNJrEdQ+P7Q0e3Zit8cPs1to8K2jZOwd8ORl2fwZnqrhCsXlfmNbENYpy35sDHl1 3AZyMCpdaKrThu1V1Wg4kJjicgz+/0BFDBuDYgjp2cF8+x7zY2qnHbQxeomF8Va7jC/Z y8oA== X-Gm-Message-State: AOAM533ZrOR/XVr+8mxTTmeULGW1eXRt4vjPqaazPQzcnw4JM1QdA+vC FxakpJx3X9FaHtE9DR8CZFnNGwPktT7jVSLj3nvYmg5054Hz5UsJd8IW+shV9SzNSmeEPRGDlj8 tuqw6HzZwlXMSbh4spQ== X-Received: by 2002:a63:484b:: with SMTP id x11mr12638687pgk.2.1615565875264; Fri, 12 Mar 2021 08:17:55 -0800 (PST) X-Google-Smtp-Source: ABdhPJxoPud6r0a0OwOeSTrz1Dcthe8LP+/uEo6txirwh7WqzuVdpWQZDjGKQDBzzIQ8yl6vQkbHsg== X-Received: by 2002:a63:484b:: with SMTP id x11mr12638653pgk.2.1615565874862; Fri, 12 Mar 2021 08:17:54 -0800 (PST) Received: from xiangao.remote.csb ([209.132.188.80]) by smtp.gmail.com with ESMTPSA id v18sm6247263pfn.117.2021.03.12.08.17.52 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 12 Mar 2021 08:17:54 -0800 (PST) Date: Sat, 13 Mar 2021 00:17:44 +0800 From: Gao Xiang To: Zorro Lang Cc: linux-xfs@vger.kernel.org, fstests@vger.kernel.org, "Darrick J. Wong" Subject: Re: [RFC PATCH v2 3/3] xfs: stress test for shrinking free space in the last AG Message-ID: <20210312161744.GB276830@xiangao.remote.csb> References: <20210312132300.259226-1-hsiangkao@redhat.com> <20210312132300.259226-4-hsiangkao@redhat.com> <20210312161755.GL3499219@localhost.localdomain> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline In-Reply-To: <20210312161755.GL3499219@localhost.localdomain> User-Agent: Mutt/1.10.1 (2018-07-13) Precedence: bulk List-ID: X-Mailing-List: fstests@vger.kernel.org On Sat, Mar 13, 2021 at 12:17:55AM +0800, Zorro Lang wrote: > On Fri, Mar 12, 2021 at 09:23:00PM +0800, Gao Xiang wrote: > > This adds a stress testcase to shrink free space as much as > > possible in the last AG with background fsstress workload. > > > > The expectation is that no crash happens with expected output. > > > > Signed-off-by: Gao Xiang > > --- > > Note that I don't use _fill_fs instead, since fill_scratch here mainly to > > eat 125M to make fsstress more effectively, rather than fill data as > > much as possible. > > As Darrick had given lots of review points to this case, I just have > 2 picky questions as below:) > > > > > tests/xfs/991 | 121 ++++++++++++++++++++++++++++++++++++++++++++++ > > tests/xfs/991.out | 8 +++ > > tests/xfs/group | 1 + > > 3 files changed, 130 insertions(+) > > create mode 100755 tests/xfs/991 > > create mode 100644 tests/xfs/991.out > > > > diff --git a/tests/xfs/991 b/tests/xfs/991 > > new file mode 100755 > > index 00000000..22a5ac81 > > --- /dev/null > > +++ b/tests/xfs/991 > > @@ -0,0 +1,121 @@ > > +#! /bin/bash > > +# SPDX-License-Identifier: GPL-2.0 > > +# Copyright (c) 2020-2021 Red Hat, Inc. All Rights Reserved. > > +# > > +# FS QA Test 991 > > +# > > +# XFS online shrinkfs stress test > > +# > > +# This test attempts to shrink unused space as much as possible with > > +# background fsstress workload. It will decrease the shrink size if > > +# larger size fails. And totally repeat 2 * TIME_FACTOR times. > > +# > > +seq=`basename $0` > > +seqres=$RESULT_DIR/$seq > > +echo "QA output created by $seq" > > + > > +here=`pwd` > > +tmp=/tmp/$$ > > +status=1 # failure is the default! > > +trap "rm -f $tmp.*; exit \$status" 0 1 2 3 15 > > + > > +# get standard environment, filters and checks > > +. ./common/rc > > +. ./common/filter > > + > > +create_scratch() > > +{ > > + _scratch_mkfs_xfs $@ | tee -a $seqres.full | \ > > + _filter_mkfs 2>$tmp.mkfs >/dev/null > > + . $tmp.mkfs > > + > > + if ! _try_scratch_mount 2>/dev/null; then > > + echo "failed to mount $SCRATCH_DEV" > > + exit 1 > > + fi > > + > > + # fix the reserve block pool to a known size so that the enospc > > + # calculations work out correctly. > > + _scratch_resvblks 1024 > /dev/null 2>&1 > > +} > > + > > +fill_scratch() > > +{ > > + $XFS_IO_PROG -f -c "resvsp 0 ${1}" $SCRATCH_MNT/resvfile > > +} > > + > > +stress_scratch() > > +{ > > + procs=3 > > + nops=$((1000 * LOAD_FACTOR)) > > + # -w ensures that the only ops are ones which cause write I/O > > + FSSTRESS_ARGS=`_scale_fsstress_args -d $SCRATCH_MNT -w -p $procs \ > > + -n $nops $FSSTRESS_AVOID` > > + $FSSTRESS_PROG $FSSTRESS_ARGS >> $seqres.full 2>&1 & > > +} > > + > > +# real QA test starts here > > +_supported_fs xfs > > +_require_scratch > > +_require_xfs_shrink > > +_require_xfs_io_command "falloc" > > Do I miss something? I only found you use xfs_io "resvsp", why you need "falloc" cmd? As I mentioned before, the testcase was derived from xfs/104 with some modification. At a quick glance, this line was added by commit 09e94f84d929 ("xfs: don't assume preallocation is always supported on XFS"). I have no more background yet. > > > + > > +rm -f $seqres.full > > +_scratch_mkfs_xfs | tee -a $seqres.full | _filter_mkfs 2>$tmp.mkfs > > +. $tmp.mkfs # extract blocksize and data size for scratch device > > + > > +decsize=`expr 42 \* 1048576` # shrink in chunks of this size at most > > +endsize=`expr 125 \* 1048576` # stop after shrinking this big > > +[ `expr $endsize / $dbsize` -lt $dblocks ] || _notrun "Scratch device too small" > > + > > +nags=2 > > +totalcount=$((2 * TIME_FACTOR)) > > + > > +while [ $totalcount -gt 0 ]; do > > + size=`expr 1010 \* 1048576` # 1010 megabytes initially > > + logblks=$(_scratch_find_xfs_min_logblocks -dsize=${size} -dagcount=${nags}) > > + > > + create_scratch -lsize=${logblks}b -dsize=${size} -dagcount=${nags} > > + > > + for i in `seq 125 -1 90`; do > > + fillsize=`expr $i \* 1048576` > > + out="$(fill_scratch $fillsize 2>&1)" > > + echo "$out" | grep -q 'No space left on device' && continue > > + test -n "${out}" && echo "$out" > > + break > > + done > > + > > + while [ $size -gt $endsize ]; do > > + stress_scratch > > + sleep 1 > > + > > + decb=`expr $decsize / $dbsize` # in data blocks > > + while [ $decb -gt 0 ]; do > > + sizeb=`expr $size / $dbsize - $decb` > > + > > + $XFS_GROWFS_PROG -D ${sizeb} $SCRATCH_MNT \ > > + >> $seqres.full 2>&1 && break > > + > > + [ $decb -gt 100 ] && decb=`expr $decb + $RANDOM % 10` > > + decb=`expr $decb / 2` > > + done > > + > > + wait > > + [ $decb -eq 0 ] && break > > + > > + # get latest dblocks > > + $XFS_INFO_PROG $SCRATCH_MNT 2>&1 | _filter_mkfs 2>$tmp.growfs >/dev/null > > + . $tmp.growfs > > + > > + size=`expr $dblocks \* $dbsize` > > + _scratch_cycle_mount > > + done > > + > > + _scratch_unmount > > + _repair_scratch_fs >> $seqres.full > > + totalcount=`expr $totalcount - 1` > > +done > > + > > +echo "*** done" > > +status=0 > > +exit > > diff --git a/tests/xfs/991.out b/tests/xfs/991.out > > new file mode 100644 > > index 00000000..e8209672 > > --- /dev/null > > +++ b/tests/xfs/991.out > > @@ -0,0 +1,8 @@ > > +QA output created by 991 > > +meta-data=DDEV isize=XXX agcount=N, agsize=XXX blks > > +data = bsize=XXX blocks=XXX, imaxpct=PCT > > + = sunit=XXX swidth=XXX, unwritten=X > > +naming =VERN bsize=XXX > > +log =LDEV bsize=XXX blocks=XXX > > +realtime =RDEV extsz=XXX blocks=XXX, rtextents=XXX > > +*** done > > diff --git a/tests/xfs/group b/tests/xfs/group > > index a7981b67..b479ed3a 100644 > > --- a/tests/xfs/group > > +++ b/tests/xfs/group > > @@ -526,3 +526,4 @@ > > 526 auto quick mkfs > > 527 auto quick quota > > 990 auto quick growfs > > +991 auto quick growfs > > Is this "stress" test case really "quick" :) ok, will update. Thanks, Gao Xiang > > > -- > > 2.27.0 > > >