From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from smtp1.linuxfoundation.org (smtp1.linux-foundation.org [172.17.192.35]) by mail.linuxfoundation.org (Postfix) with ESMTPS id 85300ACC for ; Mon, 17 Jun 2019 14:19:00 +0000 (UTC) Received: from perceval.ideasonboard.com (perceval.ideasonboard.com [213.167.242.64]) by smtp1.linuxfoundation.org (Postfix) with ESMTPS id DBDCF2C3 for ; Mon, 17 Jun 2019 14:18:58 +0000 (UTC) Date: Mon, 17 Jun 2019 17:18:39 +0300 From: Laurent Pinchart To: Mauro Carvalho Chehab Message-ID: <20190617141839.GB4777@pendragon.ideasonboard.com> References: <1559836116.15946.27.camel@HansenPartnership.com> <20190606155846.GA31044@kroah.com> <1559838275.3144.6.camel@HansenPartnership.com> <20190613105916.66d03adf@coco.lan> <20190614101222.GA4797@pendragon.ideasonboard.com> <20190614102424.3fc40f04@coco.lan> <20190614135807.GA6573@kroah.com> <20190614121137.02b8a6dc@coco.lan> <20190615110107.GA5974@pendragon.ideasonboard.com> <20190617080315.4d8fd076@coco.lan> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <20190617080315.4d8fd076@coco.lan> Cc: James Bottomley , media-submaintainers@linuxtv.org, ksummit-discuss@lists.linuxfoundation.org Subject: Re: [Ksummit-discuss] [MAINTAINERS SUMMIT] Pull network and Patch Acceptance Consistency List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Hi Mauro, On Mon, Jun 17, 2019 at 08:03:15AM -0300, Mauro Carvalho Chehab wrote: > Em Sat, 15 Jun 2019 14:01:07 +0300 Laurent Pinchart escreveu: > > On Fri, Jun 14, 2019 at 12:11:37PM -0300, Mauro Carvalho Chehab wrote: > >> Em Fri, 14 Jun 2019 15:58:07 +0200 Greg KH escreveu: > >>> On Fri, Jun 14, 2019 at 10:24:24AM -0300, Mauro Carvalho Chehab wrote: > >>>> Em Fri, 14 Jun 2019 13:12:22 +0300 Laurent Pinchart escreveu: > >>>>> On Thu, Jun 13, 2019 at 10:59:16AM -0300, Mauro Carvalho Chehab wrote: > >>>>>> Em Thu, 06 Jun 2019 19:24:35 +0300 James Bottomley escreveu: > >>>>>> > >>>>>>> [splitting issues to shorten replies] > >>>>>>> On Thu, 2019-06-06 at 17:58 +0200, Greg KH wrote: > >>>>>>>> On Thu, Jun 06, 2019 at 06:48:36PM +0300, James Bottomley wrote: > >>>>>>>>> This is probably best done as two separate topics > >>>>>>>>> > >>>>>>>>> 1) Pull network: The pull depth is effectively how many pulls your > >>>>>>>>> tree does before it goes to Linus, so pull depth 0 is sent straight > >>>>>>>>> to Linus, pull depth 1 is sent to a maintainer who sends to Linus > >>>>>>>>> and so on. We've previously spent time discussing how increasing > >>>>>>>>> the pull depth of the network would reduce the amount of time Linus > >>>>>>>>> spends handling pull requests. However, in the areas I play, like > >>>>>>>>> security, we seem to be moving in the opposite direction > >>>>>>>>> (encouraging people to go from pull depth 1 to pull depth 0). If > >>>>>>>>> we're deciding to move to a flat tree model, where everything is > >>>>>>>>> depth 0, that's fine, I just think we could do with making a formal > >>>>>>>>> decision on it so we don't waste energy encouraging greater tree > >>>>>>>>> depth. > >>>>>>>> > >>>>>>>> That depth "change" was due to the perceived problems that having a > >>>>>>>> deeper pull depth was causing. To sort that out, Linus asked for > >>>>>>>> things to go directly to him. > >>>>>>> > >>>>>>> This seems to go beyond problems with one tree and is becoming a trend. > >>>>>>> > >>>>>>>> It seems like the real issue is the problem with that subsystem > >>>>>>>> collection point, and the fact that the depth changed is a sign that > >>>>>>>> our model works well (i.e. everyone can be routed around.) > >>>>>>> > >>>>>>> I'm not really interested in calling out "problem" maintainers, or > >>>>>>> indeed having another "my patch collection method is better than yours" > >>>>>>> type discussion. What I was fishing for is whether the general > >>>>>>> impression that greater tree depth is worth striving for is actually > >>>>>>> correct, or we should all give up now and simply accept that the > >>>>>>> current flat tree is the best we can do, and, indeed is the model that > >>>>>>> works best for Linus. I get the impression this may be the case, but I > >>>>>>> think making sure by having an actual discussion among the interested > >>>>>>> parties who will be at the kernel summit, would be useful. > >>>>>> > >>>>>> On media, we came from a "depth 1" model, moving toward a "depth 2" level: > >>>>>> > >>>>>> patch author -> media/driver maintainer -> subsystem maintainer -> Linus > >>>>> > >>>>> I'd like to use this opportunity to ask again for pull requests to be > >>>>> pulled instead of cherry-picked. > >>>> > >>>> There are other forums for discussing internal media maintainership, > >>>> like the weekly meetings we have and our own mailing lists. > >>> > >>> You all have weekly meetings? That's crazy... > >> > >> Yep, every week we do a meeting, usually taking about 1 hour via irc, > >> on this channel: > >> > >> https://linuxtv.org/irc/irclogger_logs//media-maint > >> > >>> Anyway, I'll reiterate Laurent here, keeping things as a pull instead of > >>> cherry-picking does make things a lot easier for contributors. I know > >>> I'm guilty of it as well as a maintainer, but that's only until I start > >>> trusting the submitter. Once that happens, pulling is _much_ easier as > >>> a maintainer instead of individual patches for the usual reason that > >>> linux-next has already verified that the sub-tree works properly before > >>> I merge it in. > >>> > >>> Try it, it might make your load be reduced, it has for me. > >> > >> If you think this is relevant to a broader audience, let me reply with > >> a long answer about that. I prepared it and intended to reply to our > >> internal media maintainer's ML (added as c/c). > >> > >> Yet, I still think that this is media maintainer's dirty laundry > >> and should be discussed elsewhere ;-) > > > > I'll do my best to reply below with comments that are not too specific > > to the media subsystem, hoping it will be useful for a wider audience > > :-) > > > >> --- > >> > >> Laurent, > >> > >> I already explained a few times, including during the last Media Summit, > >> but it seems you missed the point. > >> > >> As shown on our stats: > >> https://linuxtv.org/patchwork_stats.php > >> > >> We're receiving about 400 to 1000 patches per month, meaning 18 to 45 > >> patches per working days (22 days/month). From those, we accept about > >> 100 to 300 patches per month (4.5 to 13.6 patches per day). > >> > >> Currently, I review all accepted patches. > > > > As other have said or hinted, this is where things start getting wrong. > > As a maintainer your duty isn't to work for 24h a day and review every > > single patch. The duty of a maintainer is to help the subsystem stay > > healthy and move forward. This can involve lots of technical work, but > > it doesn't have to, that can also be delegated (providing, of course, > > that the subsysteù would have technically competent and reliable > > contributors who would be willing to help there). In my opinion > > maintaining a subsystem is partly a technical job, and partly a social > > job. Being excellent at both is the icing on the cake, not a minimal > > requirement. > > There are a couple of reasons why I keep doing that. Among them: > > 1) I'd like to follow what's happening at the subsystem. Reviewing the > patches allow me to have at least a rough idea about what's happening, > with makes easier when we need to discuss about possible changes at > the core; I don't think anyone is calling for maintainers not to follow what is happening :-) That should however not by itself be a reason to introduce bottlenecks. > 2) An additional reviewer improves code quality. One of the feedbacks > I get from sub-maintainers is that we need more core review. So, I'm > doing my part. I agree here too, but once again, it's not a reason to introduce bottlenecks. You can certainly catch issues in patches that are submitted, and it can then improve quality of the code that gets merged, but that's true for every experienced reviewer, and we don't require review coverage of all core developers for every single patch. > 3) I like my work. > > >> I have bandwidth to review 4.5 to 13.6 patches per day, not without a lot > >> of personal efforts. For that, I use part of my spare time, as I have other > >> duties, plus I develop patches myself. So, in order to be able to handle > >> those, I typically work almost non-stop starting at 6am and sometimes > >> going up to 10pm. Also, when there are too much stuff pending (like on > >> busy months), I handle patches also during weekends. > > > > I wasn't aware of your personal work schedule, and I'm sorry to hear > > it's so extreme. This is not sustainable, and I think this clearly shows > > that a purely flat tree model with a single maintainer has difficulty > > scaling for large subsystems. If anything, this calls in my opinion for > > increasing the pull network depth to make your job bearable again. > > It has been sustainable. I've doing it over the last 10 years. I'm sorry, but the above description doesn't look at all sustainable to me, or even healthy (neither from a person's point of view nor from a subsystem's point of view). The fact that the status quo has been preserved for a long time doesn't necessarily mean it's a desirable or good option. > It is not every day I go from 6am to 10pm. Also, it is not that I don't have > a social life. I still have time for my hobbies and for my family. > > >> However, 45 patches/day (225 patches per week) is a lot for me to > >> review. I can't commit to handle such amount of patches. > >> > >> That's why I review patches after a first review from the other > >> media maintainers. The way I identify the patches I should review is > >> when I receive pull requests. > >> > >> We could do a different workflow. For example, once a media maintainer > >> review a patch, it could be delegated to me at patchwork. This would likely > >> increase the time for merging stuff, as the workflow would change from: > >> > >> +-------+ +------------------+ +---------------+ > >> | patch | -> | media maintainer | -> | submaintainer | > >> +-------+ +------------------+ +---------------+ > >> > >> to: > >> > >> +-------+ +------------------+ +---------------+ +------------------+ +---------------+ > >> | patch | -> | media maintainer | -> | submaintainer | -> | media maintainer | -> | submaintainer | > >> +-------+ +------------------+ +---------------+ +------------------+ +---------------+ > >> > >> \------------------------v--------------------------/ \---------------------v------------------/ > >> Patchwork Pull Request > >> > >> The pull request part of the new chain could eventually be (semi-)automated > >> by some scripting that would just do a checksum sum at the received patches > >> that were previously reviewed by me. If matches, and if it passes on the > >> usual checks I run for PR patches, it would push on some tree. Still, it > >> would take more time than the previous flow. > > > > I'm sorry, but I don't think this goes in the right direction. With the > > number of patches increasing, and the number of hours in a maintainer's > > day desperately not agreeing to increase above 24, the only scalable > > solution I see is to stop reviewing every single patch that is accepted > > in the subsystem tree, through delegation/sharing of maintainer's > > duties, and trust. I know it can be difficult to let go of a driver one > > has authored and let it live its life, so I can only guess the > > psychological effect is much worse for a whole subsystem. I've authored > > drivers that I cared and still care about, and I need to constantly > > remind me that too much love can lead to suffocating. The most loving > > parent has to accept that their children will one day leave home, but > > that it doesn't mean their lives will part forever. I think the same > > applies to free software. > > That's not the point. The point is that, while I have time for doing > patch reviews, I want to keep doing it. > > Also, as I said, this is media dirty laundering: weather I would keep > doing patch reviews or not - and how this will work - is something for > our internal discussions, and not for KS. As Mark said, while this discussion uses the media subsystem as an example, the best practices to handle pull networks is a kernel-wide problem. > >> Also, as also discussed during the media summit, in order to have such > >> kind of automation, we would need to improve our infrastructure, moving > >> the tests from a noisy/heated server I have over my desk to some VM > >> inside the cloud, once we get funds for it. > > > > Sure, and I think this is a topic that would gain from being discussed > > with a wider audience. The media subsystem isn't the only one to be > > large enough that it would benefit a lot from automation (I would even > > argue that all subsystems could benefit from that), so sharing > > experiences, and hearing other subsystem's wishes, would be useful here. > > Maybe. > > Are there any other subsystem currently working to get funding for > hosting/automation? The example that immediately comes to my mind is DRM/KMS (which since v4.0 has merged between 2 and 8 times as many patches as the media subsystem, depending on the kernel version, so they require lots of review bandwidth as well). With the X.org fundation, under the SPI umbrella, they raise a significant amount of money to cover the gitlab CI hosting costs (the cost comes from hosting an open-source gitlab instance, not in licenses paid to gitlab). I am personally quite fond of going through a real non-profit foundation, as it brings transparency to the accounting. > >> In any case, a discussion that affects the patch latency and our internal > >> procedures within the media subsystem is something that should be discussed > >> with other media mantainers, and not at KS. > > > > Isn't improving patch latency something that would be welcome throughout > > the kernel ? > > Your proposed change won't improve it. It will either keep the same, > if we keep the current flow: > > +-------+ +------------------+ +---------------+ > | patch | -> | media maintainer | -> | submaintainer | > +-------+ +------------------+ +---------------+ > > (either if I review the patch or not, the flow will be the same - and > so the patch latency) > > Or make it higher, if we change it to: > > +-------+ +------------------+ +---------------+ +------------------+ +---------------+ > | patch | -> | media maintainer | -> | submaintainer | -> | media maintainer | -> | submaintainer | > +-------+ +------------------+ +---------------+ +------------------+ +---------------+ > > \------------------------v--------------------------/ \---------------------v------------------/ > Patchwork Pull Request The whole idea is that it wouldn't need to go through you in the first place. Patches would be reviewed and applied by "submaintainers" (I've grown quite unhappy with that word as it can be quite belittling, but that's a separate issue) to their tree, who would then send pull requests to you. > More below: > > >> - > >> > >> That's said, one day I may not be able to review all accepted patches. > >> When this day comes, I'll just apply the pull requests I receive. > >> > >> - > >> > >> Finally, if you're so interested on improving our maintenance model, > >> I beg you: please handle the patches delegated to you: > >> > >> https://patchwork.linuxtv.org/project/linux-media/list/?series=&submitter=&state=&q=&archive=&delegate=2510 > >> > >> As we agreed on our media meetings, I handled about ~60 patches that > >> were waiting for your review since 2017 a couple of weeks ago - > >> basically the ones that are not touching the drivers you currently > >> maintain, but there are still 23 patches sent between 2013-2018 > >> over there, plus the 48 patches sent in 2019. > > Reviewing the patches on your queue or delegating them to others is > actually a concrete thing you can do in order to reduce the patch > handling latency. I've purposefully refrained from answering this part in my previous e-mail, but as you bring it up, here's what I had written and deleted. We are here reaching the interesting topic of how to motivate (or demotivate) contributors. I think it's worth a global discussion as it applies to the kernel in general (and not just to the linux-media subsystem), but I think it would overtake this discussion thread. If someone has an interest in discussing this topic, let me know and I'll split it to another thread. -- Regards, Laurent Pinchart