All of lore.kernel.org
 help / color / mirror / Atom feed
From: Ming Lei <ming.lei@redhat.com>
To: Oleksandr Natalenko <oleksandr@natalenko.name>
Cc: linux-kernel@vger.kernel.org, Jens Axboe <axboe@fb.com>,
	Christoph Hellwig <hch@lst.de>, Sagi Grimberg <sagi@grimberg.me>,
	linux-nvme@lists.infradead.org,
	David Jeffery <djeffery@redhat.com>,
	Laurence Oberman <loberman@redhat.com>,
	Paolo Valente <paolo.valente@linaro.org>, Jan Kara <jack@suse.cz>,
	Sasha Levin <sashal@kernel.org>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	Keith Busch <kbusch@kernel.org>
Subject: Re: New warning in nvme_setup_discard
Date: Wed, 21 Jul 2021 16:00:56 +0800	[thread overview]
Message-ID: <YPfUOKrxGs6FjaOZ@T590> (raw)
In-Reply-To: <2356877.Yf5hrMSTGe@natalenko.name>

On Tue, Jul 20, 2021 at 11:05:29AM +0200, Oleksandr Natalenko wrote:
> Hello, Ming.
> 
> On pondělí 19. července 2021 8:27:29 CEST Oleksandr Natalenko wrote:
> > On pondělí 19. července 2021 3:40:40 CEST Ming Lei wrote:
> > > On Sat, Jul 17, 2021 at 02:35:14PM +0200, Oleksandr Natalenko wrote:
> > > > On sobota 17. července 2021 14:19:59 CEST Oleksandr Natalenko wrote:
> > > > > On sobota 17. července 2021 14:11:05 CEST Oleksandr Natalenko wrote:
> > > > > > On sobota 17. července 2021 11:35:32 CEST Ming Lei wrote:
> > > > > > > Maybe you need to check if the build is OK, I can't reproduce it
> > > > > > > in
> > > > > > > my
> > > > > > > VM, and BFQ is still builtin:
> > > > > > > 
> > > > > > > [root@ktest-01 ~]# uname -a
> > > > > > > Linux ktest-01 5.14.0-rc1+ #52 SMP Fri Jul 16 18:56:36 CST 2021
> > > > > > > x86_64
> > > > > > > x86_64 x86_64 GNU/Linux [root@ktest-01 ~]# cat
> > > > > > > /sys/block/nvme0n1/queue/scheduler
> > > > > > > [none] mq-deadline kyber bfq
> > > > > > 
> > > > > > I don't think this is an issue with the build… BTW, with
> > > > > > `initcall_debug`:
> > > > > > 
> > > > > > ```
> > > > > > [    0.902555] calling  bfq_init+0x0/0x8b @ 1
> > > > > > [    0.903448] initcall bfq_init+0x0/0x8b returned -28 after 507
> > > > > > usecs
> > > > > > ```
> > > > > > 
> > > > > > -ENOSPC? Why? Also re-tested with the latest git tip, same result
> > > > > > :(.
> > > > > 
> > > > > OK, one extra pr_info, and I see this:
> > > > > 
> > > > > ```
> > > > > [    0.871180] blkcg_policy_register: BLKCG_MAX_POLS too small
> > > > > [    0.871612] blkcg_policy_register: -28
> > > > > ```
> > > > > 
> > > > > What does it mean please :)? The value seems to be hard-coded:
> > > > > 
> > > > > ```
> > > > > include/linux/blkdev.h
> > > > > 60:#define BLKCG_MAX_POLS               5
> > > > > ```
> > > > 
> > > > OK, after increasing this to 6 I've got my BFQ back. Please see [1].
> > > > 
> > > > [1]
> > > > https://lore.kernel.org/linux-block/20210717123328.945810-1-oleksandr@na
> > > > t
> > > > alenko.name/
> > > 
> > > OK, after you fixed the issue in blkcg_policy_register(), can you
> > > reproduce the discard issue on v5.14-rc1 with BFQ applied? If yes,
> > > can you test the patch I posted previously?
> > 
> > Yes, the issue is reproducible with both v5.13.2 and v5.14-rc1. I haven't
> > managed to reproduce it with v5.13.2+your patch. Now I will build v5.14-
> > rc2+your patch and test further.
> 
> I'm still hammering v5.14-rc2 + your patch, and I cannot reproduce the issue. 
> Given I do not have a reliable reproducer (I'm just firing up the kernel build, 
> and the issue pops up eventually, sooner or later, but usually within a couple 
> of first tries), for how long I should hammer it for your fix to be considered 
> proven?

You mentioned that the issue is reproducible with v5.14-rc, that means
it can be always reproduced in limited time(suppose it is A). If the issue
can't be reproduced any more after applying the patch in long enough time B(B >> A),
we can think it is fixed by the patch.

For example, if A is one hour, we can set B as 5*A or bigger to simulate
the long enough time.


Thanks,
Ming


WARNING: multiple messages have this Message-ID (diff)
From: Ming Lei <ming.lei@redhat.com>
To: Oleksandr Natalenko <oleksandr@natalenko.name>
Cc: linux-kernel@vger.kernel.org, Jens Axboe <axboe@fb.com>,
	Christoph Hellwig <hch@lst.de>, Sagi Grimberg <sagi@grimberg.me>,
	linux-nvme@lists.infradead.org,
	David Jeffery <djeffery@redhat.com>,
	Laurence Oberman <loberman@redhat.com>,
	Paolo Valente <paolo.valente@linaro.org>, Jan Kara <jack@suse.cz>,
	Sasha Levin <sashal@kernel.org>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	Keith Busch <kbusch@kernel.org>
Subject: Re: New warning in nvme_setup_discard
Date: Wed, 21 Jul 2021 16:00:56 +0800	[thread overview]
Message-ID: <YPfUOKrxGs6FjaOZ@T590> (raw)
In-Reply-To: <2356877.Yf5hrMSTGe@natalenko.name>

On Tue, Jul 20, 2021 at 11:05:29AM +0200, Oleksandr Natalenko wrote:
> Hello, Ming.
> 
> On pondělí 19. července 2021 8:27:29 CEST Oleksandr Natalenko wrote:
> > On pondělí 19. července 2021 3:40:40 CEST Ming Lei wrote:
> > > On Sat, Jul 17, 2021 at 02:35:14PM +0200, Oleksandr Natalenko wrote:
> > > > On sobota 17. července 2021 14:19:59 CEST Oleksandr Natalenko wrote:
> > > > > On sobota 17. července 2021 14:11:05 CEST Oleksandr Natalenko wrote:
> > > > > > On sobota 17. července 2021 11:35:32 CEST Ming Lei wrote:
> > > > > > > Maybe you need to check if the build is OK, I can't reproduce it
> > > > > > > in
> > > > > > > my
> > > > > > > VM, and BFQ is still builtin:
> > > > > > > 
> > > > > > > [root@ktest-01 ~]# uname -a
> > > > > > > Linux ktest-01 5.14.0-rc1+ #52 SMP Fri Jul 16 18:56:36 CST 2021
> > > > > > > x86_64
> > > > > > > x86_64 x86_64 GNU/Linux [root@ktest-01 ~]# cat
> > > > > > > /sys/block/nvme0n1/queue/scheduler
> > > > > > > [none] mq-deadline kyber bfq
> > > > > > 
> > > > > > I don't think this is an issue with the build… BTW, with
> > > > > > `initcall_debug`:
> > > > > > 
> > > > > > ```
> > > > > > [    0.902555] calling  bfq_init+0x0/0x8b @ 1
> > > > > > [    0.903448] initcall bfq_init+0x0/0x8b returned -28 after 507
> > > > > > usecs
> > > > > > ```
> > > > > > 
> > > > > > -ENOSPC? Why? Also re-tested with the latest git tip, same result
> > > > > > :(.
> > > > > 
> > > > > OK, one extra pr_info, and I see this:
> > > > > 
> > > > > ```
> > > > > [    0.871180] blkcg_policy_register: BLKCG_MAX_POLS too small
> > > > > [    0.871612] blkcg_policy_register: -28
> > > > > ```
> > > > > 
> > > > > What does it mean please :)? The value seems to be hard-coded:
> > > > > 
> > > > > ```
> > > > > include/linux/blkdev.h
> > > > > 60:#define BLKCG_MAX_POLS               5
> > > > > ```
> > > > 
> > > > OK, after increasing this to 6 I've got my BFQ back. Please see [1].
> > > > 
> > > > [1]
> > > > https://lore.kernel.org/linux-block/20210717123328.945810-1-oleksandr@na
> > > > t
> > > > alenko.name/
> > > 
> > > OK, after you fixed the issue in blkcg_policy_register(), can you
> > > reproduce the discard issue on v5.14-rc1 with BFQ applied? If yes,
> > > can you test the patch I posted previously?
> > 
> > Yes, the issue is reproducible with both v5.13.2 and v5.14-rc1. I haven't
> > managed to reproduce it with v5.13.2+your patch. Now I will build v5.14-
> > rc2+your patch and test further.
> 
> I'm still hammering v5.14-rc2 + your patch, and I cannot reproduce the issue. 
> Given I do not have a reliable reproducer (I'm just firing up the kernel build, 
> and the issue pops up eventually, sooner or later, but usually within a couple 
> of first tries), for how long I should hammer it for your fix to be considered 
> proven?

You mentioned that the issue is reproducible with v5.14-rc, that means
it can be always reproduced in limited time(suppose it is A). If the issue
can't be reproduced any more after applying the patch in long enough time B(B >> A),
we can think it is fixed by the patch.

For example, if A is one hour, we can set B as 5*A or bigger to simulate
the long enough time.


Thanks,
Ming


_______________________________________________
Linux-nvme mailing list
Linux-nvme@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-nvme

  reply	other threads:[~2021-07-21  8:02 UTC|newest]

Thread overview: 52+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-07-15 13:56 New warning in nvme_setup_discard Oleksandr Natalenko
2021-07-15 13:56 ` Oleksandr Natalenko
2021-07-15 14:19 ` Greg Kroah-Hartman
2021-07-15 14:19   ` Greg Kroah-Hartman
2021-07-15 14:21   ` Oleksandr Natalenko
2021-07-15 14:21     ` Oleksandr Natalenko
2021-07-15 21:37   ` Laurence Oberman
2021-07-15 21:37     ` Laurence Oberman
2021-07-16  5:50     ` Oleksandr Natalenko
2021-07-16  5:50       ` Oleksandr Natalenko
2021-07-16  2:16 ` Ming Lei
2021-07-16  2:16   ` Ming Lei
2021-07-16  5:53   ` Oleksandr Natalenko
2021-07-16  5:53     ` Oleksandr Natalenko
2021-07-16  9:33     ` Ming Lei
2021-07-16  9:33       ` Ming Lei
2021-07-16 10:03       ` Oleksandr Natalenko
2021-07-16 10:03         ` Oleksandr Natalenko
2021-07-16 10:41         ` Ming Lei
2021-07-16 10:41           ` Ming Lei
2021-07-16 12:56           ` Oleksandr Natalenko
2021-07-16 12:56             ` Oleksandr Natalenko
2021-07-17  9:35             ` Ming Lei
2021-07-17  9:35               ` Ming Lei
2021-07-17 12:11               ` Oleksandr Natalenko
2021-07-17 12:11                 ` Oleksandr Natalenko
2021-07-17 12:19                 ` Oleksandr Natalenko
2021-07-17 12:19                   ` Oleksandr Natalenko
2021-07-17 12:35                   ` Oleksandr Natalenko
2021-07-17 12:35                     ` Oleksandr Natalenko
2021-07-19  1:40                     ` Ming Lei
2021-07-19  1:40                       ` Ming Lei
2021-07-19  6:27                       ` Oleksandr Natalenko
2021-07-19  6:27                         ` Oleksandr Natalenko
2021-07-20  9:05                         ` Oleksandr Natalenko
2021-07-20  9:05                           ` Oleksandr Natalenko
2021-07-21  8:00                           ` Ming Lei [this message]
2021-07-21  8:00                             ` Ming Lei
2021-07-27 15:12                             ` Oleksandr Natalenko
2021-07-27 15:12                               ` Oleksandr Natalenko
2021-07-27 15:58                               ` Ming Lei
2021-07-27 15:58                                 ` Ming Lei
2021-07-28 13:44                                 ` Oleksandr Natalenko
2021-07-28 13:44                                   ` Oleksandr Natalenko
2021-07-28 15:53                                   ` Ming Lei
2021-07-28 15:53                                     ` Ming Lei
2021-07-28 16:38                                     ` Oleksandr Natalenko
2021-07-28 16:38                                       ` Oleksandr Natalenko
2021-07-29  3:33                                       ` Ming Lei
2021-07-29  3:33                                         ` Ming Lei
2021-07-29  9:29                                         ` Ming Lei
2021-07-29  9:29                                           ` Ming Lei

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=YPfUOKrxGs6FjaOZ@T590 \
    --to=ming.lei@redhat.com \
    --cc=axboe@fb.com \
    --cc=djeffery@redhat.com \
    --cc=gregkh@linuxfoundation.org \
    --cc=hch@lst.de \
    --cc=jack@suse.cz \
    --cc=kbusch@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-nvme@lists.infradead.org \
    --cc=loberman@redhat.com \
    --cc=oleksandr@natalenko.name \
    --cc=paolo.valente@linaro.org \
    --cc=sagi@grimberg.me \
    --cc=sashal@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.