linux-fsdevel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Christian Stroetmann <stroetmann@ontolab.com>
To: Daniel Phillips <daniel@phunq.net>
Cc: David Lang <david@lang.hm>, Dave Chinner <david@fromorbit.com>,
	linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org,
	tux3@tux3.org, Theodore Ts'o <tytso@mit.edu>
Subject: Re: Tux3 Report: How fast can we fsync?
Date: Sat, 02 May 2015 18:00:58 +0200	[thread overview]
Message-ID: <5544F4BA.5060603@ontolab.com> (raw)
In-Reply-To: <d52a170b-6c42-4294-8637-0a2bd91a29c0@phunq.net>

On the 2nd of May 2015 12:26, Daniel Phillips wrote:

Aloha everybody

> On Friday, May 1, 2015 6:07:48 PM PDT, David Lang wrote:
>> On Fri, 1 May 2015, Daniel Phillips wrote:
>>> On Friday, May 1, 2015 8:38:55 AM PDT, Dave Chinner wrote:
>>>>
>>>> Well, yes - I never claimed XFS is a general purpose filesystem.  It
>>>> is a high performance filesystem. Is is also becoming more relevant
>>>> to general purpose systems as low cost storage gains capabilities
>>>> that used to be considered the domain of high performance storage...
>>>
>>> OK. Well, Tux3 is general purpose and that means we care about single
>>> spinning disk and small systems.
>>
>> keep in mind that if you optimize only for the small systems you may 
>> not scale as well to the larger ones.
>
> Tux3 is designed to scale, and it will when the time comes. I look 
> forward to putting Shardmap through its billion file test in due 
> course. However, right now it would be wise to stay focused on basic 
> functionality suited to a workstation because volunteer devs tend to 
> have those. After that, phones are a natural direction, where hard 
> core ACID commit and really smooth file ops are particularly attractive.
>

Has anybody else a deja vu?

Nevertheless, why don't you just put your fsync, your interpretations 
(ACID, shardmap, etc.) of my things (OntoFS, file system of SASOS4Fun, 
and OntoBase), and whatever gimmicks you have in mind together into a 
prototypical file system, test it, and sent a message with a short 
description focused solely on others' and your foundational ideas and 
the technical features, a WWW address where the code can be found, and 
your test results to this mailing list without your marketing and 
self-promotion BEFORE and NOT DUE COURSE respectively NOT AFTER anybody 
else does a similar work or I am so annoyed that I implement it?

Also, if it is so general that XFS and EXT4 should adapt it, then why 
don't you help to improve these file systems?

Btw.: I have rejected my claims I made in that email, but definitely not 
given up my copyright if it is valid.

>> per the ramdisk but, possibly not as relavent as you may think. This 
>> is why it's good to test on as many different systems as you can. As 
>> you run into different types of performance you can then pick ones to 
>> keep and test all the time.
>
> I keep being surprised how well it works for things we never tested 
> before.
>
>> Single spinning disk is interesting now, but will be less interesting 
>> later. multiple spinning disks in an array of some sort is going to 
>> remain very interesting for quite a while.
>
> The way to do md well is to integrate it into the block layer like 
> Freebsd does (GEOM) and expose a richer interface for the filesystem. 
> That is how I think Tux3 should work with big iron raid. I hope to be
> able to tackle that sometime before the stars start winking out.
>
>> now, some things take a lot more work to test than others. Getting 
>> time on a system with a high performance, high capacity RAID is hard, 
>> but getting hold of an SSD from Fry's is much easier. If it's a 
>> budget item, ping me directly and I can donate one for testing (the 
>> cost of a drive is within my unallocated budget and using that to 
>> improve Linux is worthwhile)
>
> Thanks.
>
>> As I'm reading Dave's comments, he isn't attacking you the way you 
>> seem to think he is. He is pointing ot that there are problems with 
>> your data, but he's also taking a lot of time to explain what's 
>> happening (and yes, some of this is probably because your simple 
>> tests with XFS made it look so bad)
>
> I hope the lightening up trend is a trend.
>
>> the other filesystems don't use naive algortihms, they use something 
>> more complex, and while your current numbers are interesting, they 
>> are only preliminary until you add something to handle fragmentation. 
>> That can cause very significant problems.
>
> Fsync is pretty much agnostic to fragmentation, so those results are 
> unlikely to change substantially even if we happen to do a lousy job 
> on allocation policy, which I naturally consider unlikely. In fact, 
> Tux3 fsync is going to get faster over time for a couple of reasons: 
> the minimum blocks per commit will be reduced, and we will get rid of 
> most of the seeks to beginning of volume that we currently suffer per 
> commit.
>
>> Remember how fabulous btrfs looked in the initial reports? and then 
>> corner cases were found that caused real problems and as the 
>> algorithms have been changed to prevent those corner cases from being 
>> so easy to hit, the common case has suffered somewhat. This isn't an 
>> attack on Tux2 or btrfs, it's just a reality of programming. If you 
>> are not accounting for all the corner cases, everything is easier, 
>> and faster.
>
>>> Mine is a lame i5 minitower with 4GB from Fry's. Yours is clearly way
>>> more substantial, so I can't compare my numbers directly to yours.
>>
>> If you are doing tests with a 4G ramdisk on a machine with only 4G of 
>> RAM, it seems like you end up testing a lot more than just the 
>> filesystem. Testing in such low memory situations can indentify 
>> significant issues, but it is questionable as a 'which filesystem is 
>> better' benchmark.
>
> A 1.3 GB tmpfs, and sorry, it is 10 GB (the machine next to it is 4G). 
> I am careful to ensure the test environment does not have spurious 
> memory or cpu hogs. I will not claim that this is the most sterile 
> test environment possible, but it is adequate for the task at hand. 
> Nearly always, when I find big variations in the test numbers it turns 
> out to be a quirk of one filesystem that is not exhibited by the 
> others. Everything gets multiple runs and lands in a spreadsheet. Any 
> fishy variance is investigated.
>
> By the way, the low variance kings by far are Ext4 and Tux3, and of 
> those two, guess which one is more consistent. XFS is usually steady, 
> but can get "emotional" with lots of tasks, and Btrfs has regular wild 
> mood swings whenever the stars change alignment. And while I'm making 
> gross generalizations: XFS and Btrfs go OOM way before Ext4 and Tux3.
>
>> Just a suggestion, but before you do a huge post about how great your 
>> filesystem is performing, making the code avaialble so that others 
>> can test it when prompted by your post is probably a very good idea. 
>> If it means that you have to send out your post a week later, it's a 
>> very small cost for the benefit of having other people able to easily 
>> try it on hardware that you don't have access to.
>
> Next time. This time I wanted it off my plate as soon as possible so I 
> could move on to enospc work. And this way is more involving, we get a 
> little suspense before the rematch.
>
>> If there is a reason to post wihtout the code being in the main, 
>> publicised repo, then your post should point people at what code they 
>> can use to duplicate it.
>
> I could have included the patch in the post, it is small enough. If it 
> still isn't in the repo in a few days then I will post it, to avoid 
> giving the impression I'm desperately trying to fix obscure bugs in 
> it, which isn't the case.
>
>> but really, 11 months without updating the main repo?? This is Open 
>> Source development, publish early and often.
>
> It's not as bad as that:
>
>   https://github.com/OGAWAHirofumi/linux-tux3/commits/hirofumi
>   https://github.com/OGAWAHirofumi/linux-tux3/commits/hirofumi-user
>
>> something to investigate, but I have seen probelms on ext* in the 
>> past. ext4 may have fixed this, or it may just have moved the point 
>> where it triggers.
>
> My spectrum of tests is small and I am not hunting for anomalies, only 
> reporting what happened to come up. It is not very surprising that some
> odd things happen with 10,000 tasks, there is probably not much test 
> coverage there. On the whole I was surprised and impressed when all 
> filesystems mostly just worked. I was expecting to hit scheduler 
> issues for one thing, and nothing obvious came up. Also, not one oops 
> on any filesystem (even Tux3) and only one assert, already reported 
> upstream and turned out to be fixed a week or two ago.
>
>>>> ...
>>> Your machine makes mine look like a PCjr. ...
>>
>> The interesting thing here is that on the faster machine btrfs didn't 
>> speed up significantly while ext4 and xfs did. It will be interesting 
>> to see what the results are for tux3
>
> The numbers are well into the something-is-really-wrong zone (and I 
> should have flagged that earlier but it was a long day). That test is 
> supposed to be -s, all synchronous, and his numbers are more typical of
> async. Needs double checking all round, including here. Anybody can 
> replicate that test, it is only an apt-get install dbench away (hint 
> hint).
>
> Differences: my numbers are kvm with loopback mount on tmpfs. His are 
> on ramdisk and probably native. I have to reboot to make a ramdisk big 
> enough to run dbench and I would rather not right now.
>
> How important is it to get to the bottom of the variance in test 
> results running on RAM? Probably important in the long run, because 
> storage devices are looking more like RAM all the time, but as of 
> today, maybe not very urgent.
>
> Also, I was half expecting somebody to question the wisdom of running 
> benchmarks under KVM instead of native, but nobody did. Just for the 
> record, I would respond: running virtual probably accounts for the
> majority of server instances today.
>
>> and both of you need to remember that while servers are getting 
>> faster, we are also seeing much lower power, weaker servers showing 
>> up as well. And while these smaller servers are not trying to do teh 
>> 10000 thread fsync workload, they are using flash based storage more 
>> frequently than they are spinning rust (frequently through the 
>> bottleneck of a SD card) so continuing tests on low end devices is good.
>
> Low end servers and embedded concerns me more, indeed.
>> what drives are available now? see if you can get a couple (either 
>> directly or donated)
>
> Right, time to hammer on flash.
>
> Regards,
>
> Daniel
> -- 

Thanks
Best Regards
Have fun
C.S.

  reply	other threads:[~2015-05-02 16:00 UTC|newest]

Thread overview: 160+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-04-28 23:13 Tux3 Report: How fast can we fsync? Daniel Phillips
2015-04-29  2:21 ` Mike Galbraith
2015-04-29  6:01   ` Daniel Phillips
2015-04-29  6:20     ` Richard Weinberger
2015-04-29  6:56       ` Daniel Phillips
2015-04-29  6:33     ` Mike Galbraith
2015-04-29  7:23       ` Daniel Phillips
2015-04-29 16:42         ` Mike Galbraith
2015-04-29 19:05           ` xfs: does mkfs.xfs require fancy switches to get decent performance? (was Tux3 Report: How fast can we fsync?) Mike Galbraith
2015-04-29 19:20             ` Austin S Hemmelgarn
2015-04-29 21:12             ` Daniel Phillips
2015-04-30  4:40               ` Mike Galbraith
2015-04-30  0:20             ` Dave Chinner
2015-04-30  3:35               ` Mike Galbraith
2015-04-30  9:00               ` Martin Steigerwald
2015-04-30 14:57                 ` Theodore Ts'o
2015-04-30 15:59                   ` Daniel Phillips
2015-04-30 17:59                   ` Martin Steigerwald
2015-04-30 11:14               ` Daniel Phillips
2015-04-30 12:07                 ` Mike Galbraith
2015-04-30 12:58                   ` Daniel Phillips
2015-04-30 13:48                     ` Mike Galbraith
2015-04-30 14:07                       ` Daniel Phillips
2015-04-30 14:28                         ` Howard Chu
2015-04-30 15:14                           ` Daniel Phillips
2015-04-30 16:00                             ` Howard Chu
2015-04-30 18:22                             ` Christian Stroetmann
2015-05-11 22:12                             ` Pavel Machek
2015-05-11 23:17                               ` Theodore Ts'o
2015-05-12  2:34                                 ` Daniel Phillips
2015-05-12  5:38                                   ` Dave Chinner
2015-05-12  6:18                                     ` Daniel Phillips
2015-05-12 18:39                                       ` David Lang
2015-05-12 20:54                                         ` Daniel Phillips
2015-05-12 21:30                                           ` David Lang
2015-05-12 22:27                                             ` Daniel Phillips
2015-05-12 22:35                                               ` David Lang
2015-05-12 23:55                                                 ` Theodore Ts'o
2015-05-13  1:26                                                 ` Daniel Phillips
2015-05-13 19:09                                                   ` Martin Steigerwald
2015-05-13 19:37                                                     ` Daniel Phillips
2015-05-13 20:02                                                       ` Jeremy Allison
2015-05-13 20:24                                                         ` Daniel Phillips
2015-05-13 20:25                                                       ` Martin Steigerwald
2015-05-13 20:38                                                         ` Daniel Phillips
2015-05-13 21:10                                                           ` Martin Steigerwald
2015-05-13  0:31                                             ` Daniel Phillips
2015-05-12 21:30                                           ` Christian Stroetmann
2015-05-13  7:20                                           ` Pavel Machek
2015-05-13 13:47                                             ` Elifarley Callado Coelho Cruz
2015-05-12  9:03                                   ` Pavel Machek
2015-05-12 11:22                                     ` Daniel Phillips
2015-05-12 13:26                                       ` Howard Chu
2015-05-11 23:53                               ` Daniel Phillips
2015-05-12  0:12                                 ` David Lang
2015-05-12  4:36                                   ` Daniel Phillips
2015-05-12 17:30                                     ` Christian Stroetmann
2015-05-13  7:25                                 ` Pavel Machek
2015-05-13 11:31                                   ` Daniel Phillips
2015-05-13 12:41                                     ` Daniel Phillips
2015-05-13 13:08                                     ` Mike Galbraith
2015-05-13 13:15                                       ` Daniel Phillips
2015-04-30 14:33                         ` Mike Galbraith
2015-04-30 15:24                           ` Daniel Phillips
2015-04-29 20:40           ` Tux3 Report: How fast can we fsync? Daniel Phillips
2015-04-29 22:06             ` OGAWA Hirofumi
2015-04-30  3:57               ` Mike Galbraith
2015-04-30  3:50             ` Mike Galbraith
2015-04-30 10:59               ` Daniel Phillips
2015-04-30  1:46 ` Dave Chinner
2015-04-30 10:28   ` Daniel Phillips
2015-05-01 15:38     ` Dave Chinner
2015-05-01 23:20       ` Daniel Phillips
2015-05-02  1:07         ` David Lang
2015-05-02 10:26           ` Daniel Phillips
2015-05-02 16:00             ` Christian Stroetmann [this message]
2015-05-02 16:30               ` Richard Weinberger
2015-05-02 17:00                 ` Christian Stroetmann
2015-05-12 17:41 ` Daniel Phillips
2015-05-12 17:46 ` Tux3 Report: How fast can we fail? Daniel Phillips
2015-05-13 22:07   ` Daniel Phillips
2015-05-26 10:03   ` Pavel Machek
2015-05-27  6:41     ` Mosis Tembo
2015-05-27 18:28       ` Daniel Phillips
2015-05-27 21:39         ` Pavel Machek
2015-05-27 22:46           ` Daniel Phillips
2015-05-28 12:55             ` Austin S Hemmelgarn
2015-05-27  7:37     ` Mosis Tembo
2015-05-27 14:04       ` Austin S Hemmelgarn
2015-05-27 15:21         ` Mosis Tembo
2015-05-27 15:37           ` Austin S Hemmelgarn
2015-05-14  7:37 ` [WIP] tux3: Optimized fsync Daniel Phillips
2015-05-14  8:26 ` [FYI] tux3: Core changes Daniel Phillips
2015-05-14 12:59   ` Rik van Riel
2015-05-15  0:06     ` Daniel Phillips
2015-05-15  3:06       ` Rik van Riel
2015-05-15  8:09         ` Mel Gorman
2015-05-15  9:54           ` Daniel Phillips
2015-05-15 11:00             ` Mel Gorman
2015-05-16 22:38               ` David Lang
2015-05-18 12:57                 ` Mel Gorman
2015-05-15  9:38         ` Daniel Phillips
2015-05-27  7:41           ` Pavel Machek
2015-05-27 18:09             ` Daniel Phillips
2015-05-27 21:37               ` Pavel Machek
2015-05-27 22:33                 ` Daniel Phillips
2015-05-15  8:05       ` Mel Gorman
2015-05-17 13:26     ` Boaz Harrosh
2015-05-18  2:20       ` Rik van Riel
2015-05-18  7:58         ` Boaz Harrosh
2015-05-19  4:46         ` Daniel Phillips
2015-05-21 19:43     ` [WIP][PATCH] tux3: preliminatry nospace handling Daniel Phillips
2015-05-19 14:00   ` [FYI] tux3: Core changes Jan Kara
2015-05-19 19:18     ` Daniel Phillips
2015-05-19 20:33       ` David Lang
2015-05-20 14:44         ` Jan Kara
2015-05-20 16:22           ` Daniel Phillips
2015-05-20 18:01             ` David Lang
2015-05-20 19:53             ` Rik van Riel
2015-05-20 22:51               ` Daniel Phillips
2015-05-21  3:24                 ` Daniel Phillips
2015-05-21  3:51                   ` David Lang
2015-05-21 19:53                     ` Daniel Phillips
2015-05-26  4:25                       ` Rik van Riel
2015-05-26  4:30                         ` Daniel Phillips
2015-05-26  6:04                           ` David Lang
2015-05-26  6:11                             ` Daniel Phillips
2015-05-26  6:13                               ` David Lang
2015-05-26  8:09                                 ` Daniel Phillips
2015-05-26 10:13                                   ` Pavel Machek
2015-05-26  7:09                               ` Jan Kara
2015-05-26  8:08                                 ` Daniel Phillips
2015-05-26  9:00                                   ` Jan Kara
2015-05-26 20:22                                     ` Daniel Phillips
2015-05-26 21:36                                       ` Rik van Riel
2015-05-26 21:49                                         ` Daniel Phillips
2015-05-27  8:41                                       ` Jan Kara
2015-06-21 15:36                                         ` OGAWA Hirofumi
2015-06-23 16:12                                           ` Jan Kara
2015-07-05 12:54                                             ` OGAWA Hirofumi
2015-07-09 16:05                                               ` Jan Kara
2015-07-31  4:44                                                 ` OGAWA Hirofumi
2015-07-31 15:37                                                   ` Raymond Jennings
2015-07-31 17:27                                                     ` Daniel Phillips
2015-07-31 18:29                                                       ` David Lang
2015-07-31 18:43                                                         ` Daniel Phillips
2015-07-31 22:12                                                         ` Daniel Phillips
2015-07-31 22:27                                                           ` David Lang
2015-08-01  0:00                                                             ` Daniel Phillips
2015-08-01  0:16                                                               ` Daniel Phillips
2015-08-03 13:07                                                                 ` Jan Kara
2015-08-01 10:55                                                             ` Elifarley Callado Coelho Cruz
2015-08-18 16:39                                                       ` Rik van Riel
2015-08-03 13:42                                                   ` Jan Kara
2015-08-09 13:42                                                     ` OGAWA Hirofumi
2015-08-10 12:45                                                       ` Jan Kara
2015-08-16 19:42                                                         ` OGAWA Hirofumi
2015-05-26 10:22                                   ` Sergey Senozhatsky
2015-05-26 12:33                                     ` Jan Kara
2015-05-26 19:18                                     ` Daniel Phillips

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=5544F4BA.5060603@ontolab.com \
    --to=stroetmann@ontolab.com \
    --cc=daniel@phunq.net \
    --cc=david@fromorbit.com \
    --cc=david@lang.hm \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=tux3@tux3.org \
    --cc=tytso@mit.edu \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).