From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.8 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E343AC433DB for ; Fri, 8 Jan 2021 08:32:07 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 92076235DD for ; Fri, 8 Jan 2021 08:32:07 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727570AbhAHIbq convert rfc822-to-8bit (ORCPT ); Fri, 8 Jan 2021 03:31:46 -0500 Received: from youngberry.canonical.com ([91.189.89.112]:39436 "EHLO youngberry.canonical.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727286AbhAHIbq (ORCPT ); Fri, 8 Jan 2021 03:31:46 -0500 Received: from mail-oo1-f70.google.com ([209.85.161.70]) by youngberry.canonical.com with esmtps (TLS1.2:ECDHE_RSA_AES_128_GCM_SHA256:128) (Exim 4.86_2) (envelope-from ) id 1kxnAj-0005KC-Fy for linux-kernel@vger.kernel.org; Fri, 08 Jan 2021 08:31:01 +0000 Received: by mail-oo1-f70.google.com with SMTP id w3so6242323oov.16 for ; Fri, 08 Jan 2021 00:31:01 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc:content-transfer-encoding; bh=bGk5OzoF1zoyQLWMK8Zm9vZAw7rYyp6h8IS5Niras3M=; b=sKKIZfxJdgl4xt1PldFjZuF+0/i7EFzcv1INYPVc81RfaoCLZ0WpP8VheTqSLn8QCx ClqDyLM/UJ5252/QiGgQWIpIAHF8tEA2CyghP9G6+5bcA2JuiEh9L65mKe03ykwmg6aC /roOz9kXrtjC5J3LmwNRWxjXmR3ceu6VIhnjJ7qkdjmp8ZawBEWNyorZ5JhLonMn7jbx 4OVN5VPlp+bKqir0bTR1Xquyq5OntpnOWHEitCIvwcZRlIBJugjn1XTAI9rSWgk/F0t+ ryYpeTfhM+sLT7XsumcZIXmIF4Ft2Bjc3+CgoH3A8Tdu7PFx2IL4kF9hu4oRUDrMawZb Bzng== X-Gm-Message-State: AOAM532njpoIAW7iiIJzKBtM+05K77ENVtzJFrdkGXqP6GN3Y1JXC6+b AO0Trq0F/FVj3i0ybR/r6w+Zj8rD1l5GFtBHb0X8oe18W9IOLAxXL+W5ShTgoRzqkaQvP3Ar7Ca ywueGFMgQAxn/I/Mn9G0eIquAMHgQ4lNzdkVSEg/JLna5CKAYIlpo7OOnhg== X-Received: by 2002:a9d:620f:: with SMTP id g15mr1737636otj.361.1610094660238; Fri, 08 Jan 2021 00:31:00 -0800 (PST) X-Google-Smtp-Source: ABdhPJzTDzveyi8ScS3wI/VLNY7NAPC5M6RuCm2S8BtQW1KWrRSmU7glLl+7kvhYU+tnE16BoLSn9L8FvorzZU86ixI= X-Received: by 2002:a9d:620f:: with SMTP id g15mr1737624otj.361.1610094659892; Fri, 08 Jan 2021 00:30:59 -0800 (PST) MIME-Version: 1.0 References: <20210105030602.14427-1-tdd21151186@gmail.com> <1a4b2a68-a7b0-8eb0-e60b-c3cf5a5a9e56@suse.de> <084276ab-7c74-31be-b957-3b039d7061a1@suse.de> In-Reply-To: <084276ab-7c74-31be-b957-3b039d7061a1@suse.de> From: Dongdong Tao Date: Fri, 8 Jan 2021 16:30:48 +0800 Message-ID: Subject: Re: [PATCH] bcache: consider the fragmentation when update the writeback rate To: Coly Li Cc: Kent Overstreet , "open list:BCACHE (BLOCK LAYER CACHE)" , open list , Gavin Guo , Gerald Yang , Trent Lloyd , Dominique Poulain , Dongsheng Yang Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: 8BIT Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Coly, They are captured with the same time length, the meaning of the timestamp and the time unit on the x-axis are different. (Sorry, I should have clarified this right after the chart) For the latency chart: The timestamp is the relative time since the beginning of the benchmark, so the start timestamp is 0 and the unit is based on millisecond For the dirty data and cache available percent chart: The timestamp is the UNIX timestamp, the time unit is based on second, I capture the stats every 5 seconds with the below script: --- #!/bin/sh while true; do echo "`date +%s`, `cat /sys/block/bcache0/bcache/dirty_data`, `cat /sys/block/bcache0/bcache/cache/cache_available_percent`, `cat /sys/block/bcache0/bcache/writeback_rate`" >> $1; sleep 5; done; --- Unfortunately, I can't easily make them using the same timestamp, but I guess I can try to convert the UNIX timestamp to the relative time like the first one. But If we ignore the value of the X-axis, we can still roughly compare them by using the length of the X-axis since they have the same time length, and we can see that the Master's write start hitting the backing device when the cache_available_percent dropped to around 30. Regards, Dongdong On Fri, Jan 8, 2021 at 12:06 PM Coly Li wrote: > > On 1/7/21 10:55 PM, Dongdong Tao wrote: > > Hi Coly, > > > > > > Thanks for the reminder, I understand that the rate is only a hint of > > the throughput, it’s a value to calculate the sleep time between each > > round of keys writeback, the higher the rate, the shorter the sleep > > time, most of the time this means the more dirty keys it can writeback > > in a certain amount of time before the hard disk running out of speed. > > > > > > Here is the testing data that run on a 400GB NVME + 1TB NVME HDD > > > > Hi Dongdong, > > Nice charts :-) > > > Steps: > > > > 1. > > > > make-bcache -B -C --writeback > > > > 2. > > > > sudo fio --name=random-writers --filename=/dev/bcache0 > > --ioengine=libaio --iodepth=1 --rw=randrw --blocksize=64k,8k > > --direct=1 --numjobs=1 --write_lat_log=mix --log_avg_msec=10 > > > The fio benchmark commands ran for about 20 hours. > > > > The time lengths of first 3 charts are 7.000e+7, rested are 1.60930e+9. > I guess the time length of the I/O latency chart is 1/100 of the rested. > > Can you also post the latency charts for 1.60930e+9 seconds? Then I can > compare the latency with dirty data and available cache charts. > > > Thanks. > > > Coly Li > > > > > > > > > Let’s have a look at the write latency first: > > > > Master: > > > > > > > > Master+the patch: > > > > Combine them together: > > > > Again, the latency (y-axis) is based on nano-second, x-axis is the > > timestamp based on milli-second, as we can see the master latency is > > obviously much higher than the one with my patch when the master bcache > > hit the cutoff writeback sync, the master isn’t going to get out of this > > cutoff writeback sync situation, This graph showed it already stuck at > > the cutoff writeback sync for about 4 hours before I finish the testing, > > it may still needs to stuck for days before it can get out this > > situation itself. > > > > > > Note that there are 1 million points for each , red represents master, > > green represents mater+my patch. Most of them are overlapped with each > > other, so it may look like this graph has more red points then green > > after it hitting the cutoff, but simply it’s because the latency has > > scaled to a bigger range which represents the HDD latency. > > > > > > > > Let’s also have a look at the bcache’s cache available percent and dirty > > data percent. > > > > Master: > > > > Master+this patch: > > > > As you can see, this patch can avoid it hitting the cutoff writeback sync. > > > > > > As to say the improvement for this patch against the first one, let’s > > take a look at the writeback rate changing during the run. > > > > patch V1: > > > > > > > > Patch V2: > > > > > > The Y-axis is the value of rate, the V1 is very aggressive as it jumps > > instantly from a minimum 8 to around 10 million. And the patch V2 can > > control the rate under 5000 during the run, and after the first round of > > writeback, it can stay even under 2500, so this proves we don’t need to > > be as aggressive as V1 to get out of the high fragment situation which > > eventually causes all writes hitting the backing device. This looks very > > reasonable for me now. > > > > Note that the fio command that I used is consuming the bucket quite > > aggressively, so it had to hit the third stage which has the highest > > aggressiveness, but I believe this is not true in a real production env, > > real production env won’t consume buckets that aggressively, so I expect > > stage 3 may not very often be needed to hit. > > > > > > As discussed, I'll run multiple block size testing on at least 1TB NVME > > device later. > > But it might take some time. > > > > > > Regards, > > Dongdong > > > > On Tue, Jan 5, 2021 at 12:33 PM Coly Li > > wrote: > > > > On 1/5/21 11:44 AM, Dongdong Tao wrote: > > > Hey Coly, > > > > > > This is the second version of the patch, please allow me to explain a > > > bit for this patch: > > > > > > We accelerate the rate in 3 stages with different aggressiveness, the > > > first stage starts when dirty buckets percent reach above > > > BCH_WRITEBACK_FRAGMENT_THRESHOLD_LOW(50), the second is > > > BCH_WRITEBACK_FRAGMENT_THRESHOLD_MID(57) and the third is > > > BCH_WRITEBACK_FRAGMENT_THRESHOLD_HIGH(64). By default the first stage > > > tries to writeback the amount of dirty data in one bucket (on average) > > > in (1 / (dirty_buckets_percent - 50)) second, the second stage > > tries to > > > writeback the amount of dirty data in one bucket in (1 / > > > (dirty_buckets_percent - 57)) * 200 millisecond. The third stage tries > > > to writeback the amount of dirty data in one bucket in (1 / > > > (dirty_buckets_percent - 64)) * 20 millisecond. > > > > > > As we can see, there are two writeback aggressiveness increasing > > > strategies, one strategy is with the increasing of the stage, the > > first > > > stage is the easy-going phase whose initial rate is trying to > > write back > > > dirty data of one bucket in 1 second, the second stage is a bit more > > > aggressive, the initial rate tries to writeback the dirty data of one > > > bucket in 200 ms, the last stage is even more, whose initial rate > > tries > > > to writeback the dirty data of one bucket in 20 ms. This makes sense, > > > one reason is that if the preceding stage couldn’t get the > > fragmentation > > > to a fine stage, then the next stage should increase the > > aggressiveness > > > properly, also it is because the later stage is closer to the > > > bch_cutoff_writeback_sync. Another aggressiveness increasing > > strategy is > > > with the increasing of dirty bucket percent within each stage, the > > first > > > strategy controls the initial writeback rate of each stage, while this > > > one increases the rate based on the initial rate, which is > > initial_rate > > > * (dirty bucket percent - BCH_WRITEBACK_FRAGMENT_THRESHOLD_X). > > > > > > The initial rate can be controlled by 3 parameters > > > writeback_rate_fp_term_low, writeback_rate_fp_term_mid, > > > writeback_rate_fp_term_high, they are default 1, 5, 50, users can > > adjust > > > them based on their needs. > > > > > > The reason that I choose 50, 57, 64 as the threshold value is because > > > the GC must be triggered at least once during each stage due to the > > > “sectors_to_gc” being set to 1/16 (6.25 %) of the total cache > > size. So, > > > the hope is that the first and second stage can get us back to good > > > shape in most situations by smoothly writing back the dirty data > > without > > > giving too much stress to the backing devices, but it might still > > enter > > > the third stage if the bucket consumption is very aggressive. > > > > > > This patch use (dirty / dirty_buckets) * fp_term to calculate the > > rate, > > > this formula means that we want to writeback (dirty / > > dirty_buckets) in > > > 1/fp_term second, fp_term is calculated by above aggressiveness > > > controller, “dirty” is the current dirty sectors, “dirty_buckets” > > is the > > > current dirty buckets, so (dirty / dirty_buckets) means the average > > > dirty sectors in one bucket, the value is between 0 to 1024 for the > > > default setting, so this formula basically gives a hint that to > > reclaim > > > one bucket in 1/fp_term second. By using this semantic, we can have a > > > lower writeback rate when the amount of dirty data is decreasing and > > > overcome the fact that dirty buckets number is always increasing > > unless > > > GC happens. > > > > > > *Compare to the first patch: > > > *The first patch is trying to write back all the data in 40 seconds, > > > this will result in a very high writeback rate when the amount of > > dirty > > > data is big, this is mostly true for the large cache devices. The > > basic > > > problem is that the semantic of this patch is not ideal, because we > > > don’t really need to writeback all dirty data in order to solve this > > > issue, and the instant large increase of the rate is something I > > feel we > > > should better avoid (I like things to be smoothly changed unless no > > > choice: )). > > > > > > Before I get to this new patch(which I believe should be optimal > > for me > > > atm), there have been many tuning/testing iterations, eg. I’ve > > tried to > > > tune the algorithm to writeback ⅓ of the dirty data in a certain > > amount > > > of seconds, writeback 1/fragment of the dirty data in a certain amount > > > of seconds, writeback all the dirty data only in those error_buckets > > > (error buckets = dirty buckets - 50% of the total buckets) in a > > certain > > > amount of time. However, those all turn out not to be ideal, only the > > > semantic of the patch makes much sense for me and allows me to control > > > the rate in a more precise way. > > > > > > *Testing data: > > > *I'll provide the visualized testing data in the next couple of days > > > with 1TB NVME devices cache but with HDD as backing device since it's > > > what we mostly used in production env. > > > I have the data for 400GB NVME, let me prepare it and take it for > > you to > > > review. > > [snipped] > > > > Hi Dongdong, > > > > Thanks for the update and continuous effort on this idea. > > > > Please keep in mind the writeback rate is just a advice rate for the > > writeback throughput, in real workload changing the writeback rate > > number does not change writeback throughput obviously. > > > > Currently I feel this is an interesting and promising idea for your > > patch, but I am not able to say whether it may take effect in real > > workload, so we do need convinced performance data on real workload and > > configuration. > > > > Of course I may also help on the benchmark, but my to-do list is long > > enough and it may take a very long delay time. > > > > Thanks. > > > > Coly Li > > >