linux-pci.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Leon Romanovsky <leon@kernel.org>
To: Alex Williamson <alex.williamson@redhat.com>
Cc: Don Dutile <ddutile@redhat.com>,
	Bjorn Helgaas <helgaas@kernel.org>,
	Bjorn Helgaas <bhelgaas@google.com>,
	Saeed Mahameed <saeedm@nvidia.com>,
	Jason Gunthorpe <jgg@nvidia.com>,
	Jakub Kicinski <kuba@kernel.org>,
	linux-pci@vger.kernel.org, linux-rdma@vger.kernel.org,
	netdev@vger.kernel.org
Subject: Re: [PATCH mlx5-next 1/4] PCI: Configure number of MSI-X vectors for SR-IOV VFs
Date: Sun, 10 Jan 2021 10:47:35 +0200	[thread overview]
Message-ID: <20210110084735.GH31158@unreal> (raw)
In-Reply-To: <20210108092145.7c70ff74@omen.home>

On Fri, Jan 08, 2021 at 09:21:45AM -0700, Alex Williamson wrote:
> On Fri, 8 Jan 2021 09:25:25 +0200
> Leon Romanovsky <leon@kernel.org> wrote:
>
> > On Thu, Jan 07, 2021 at 10:54:38PM -0500, Don Dutile wrote:
> > > On 1/7/21 7:57 PM, Bjorn Helgaas wrote:
> > > > [+cc Alex, Don]
> >
> > <...>
> >
> > > > Help me connect the dots here.  Is this required because of something
> > > > peculiar to mlx5, or is something like this required for all SR-IOV
> > > > devices because of the way the PCIe spec is written?
> > > So, overall, I'm guessing the mlx5 device can have 1000's of MSIX -- say, one per send/receive/completion queue.
> > > This device capability may exceed the max number MSIX a VM can have/support (depending on guestos).
> > > So, a sysfs tunable is used to set the max MSIX available, and thus, the device puts >1 send/rcv/completion queue intr on a given MSIX.
> > >
> > > ok, time for Leon to better state what this patch does,
> > > and why it's needed on mlx5 (and may be applicable to other/future high-MSIX devices assigned to VMs (NVME?)).
> > > Hmmm, now that I said it, why is it SRIOV-centric and not pci-device centric (can pass a PF w/high number of MSIX to a VM).
> >
> > Thanks Don and Bjorn,
> >
> > I will answer on all comments a little bit later when I will return
> > to the office (Sunday).
> >
> > However it is important for me to present the use case.
> >
> > Our mlx5 SR-IOV devices were always capable to drive many MSI-X (upto 2K,
> > don't catch me on exact number), however when user created VFs, the FW has
> > no knowledge of how those VFs will be used. So FW had no choice but statically
> > and equally assign same amount of MSI-X to all VFs.
> >
> > After SR-IOV VF creation, user will bind those new VFs to the VMs, but
> > the VMs have different number of CPUs and despite HW being able to deliver
> > all needed number of vectors (in mlx5 netdev world, number of channels == number
> > of CPUs == number of vectors), we will be limited by already set low number
> > of vectors.
> >
> > So it is not for vector reduction, but more for vector re-partition.
> >
> > As an example, imagine mlx5 with two VFs. One VF is bounded to VM with 200 CPUs
> > and another is bounded to VM with 1 CPU. They need different amount of MSI-X vectors.
> >
> > Hope that I succeeded to explain :).
>
> The idea is not unreasonable imo, but without knowing the size of the
> vector pool, range available per vf, or ultimately whether the vf
> supports this feature before we try to configure it, I don't see how
> userspace is expected to make use of this in the general case.  If the
> configuration requires such specific vf vector usage and pf driver
> specific knowledge, I'm not sure it's fit as a generic pci-sysfs
> interface.  Thanks,

I didn't prohibit read of newly created sysfs file, but if I change
the implementation to vf_msix_vec_show() to return -EOPNOTSUPP for
not-supported device, the software will be able to distinguish
supported/not-supported.

SW will read this file:
	-> success -> feature supported
	-> failure -> feature not supported

There is one extra sysfs file needed: vf_total_msix. That file will
give total number of MSI-X vectors that is possible to configure.

The same logic (supported/not-supported) can be applicable here as well.

The feature itself will be used by orchestration software that will
make decisions based on already configured values or future promises
and the overall total number. The positive outcome of this scheme
that driver stays lean.

Thanks

>
> Alex
>

  reply	other threads:[~2021-01-10  8:48 UTC|newest]

Thread overview: 16+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-01-03  8:24 [PATCH mlx5-next 0/4] Dynamically assign MSI-X vectors count Leon Romanovsky
2021-01-03  8:24 ` [PATCH mlx5-next 1/4] PCI: Configure number of MSI-X vectors for SR-IOV VFs Leon Romanovsky
2021-01-08  0:57   ` Bjorn Helgaas
2021-01-08  3:54     ` Don Dutile
2021-01-08  7:25       ` Leon Romanovsky
2021-01-08 16:21         ` Alex Williamson
2021-01-10  8:47           ` Leon Romanovsky [this message]
2021-01-08 21:09       ` Bjorn Helgaas
2021-01-09  2:54         ` Don Dutile
2021-01-10  8:33           ` Leon Romanovsky
2021-01-10  8:25         ` Leon Romanovsky
2021-01-10  8:29       ` Leon Romanovsky
2021-01-10  8:22     ` Leon Romanovsky
2021-01-03  8:24 ` [PATCH mlx5-next 3/4] net/mlx5: Dynamically assign MSI-X vectors count Leon Romanovsky
2021-01-03  8:24 ` [PATCH mlx5-next 4/4] net/mlx5: Allow to the users to configure number of MSI-X vectors Leon Romanovsky
2021-01-06  5:50 ` [PATCH mlx5-next 0/4] Dynamically assign MSI-X vectors count Leon Romanovsky

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20210110084735.GH31158@unreal \
    --to=leon@kernel.org \
    --cc=alex.williamson@redhat.com \
    --cc=bhelgaas@google.com \
    --cc=ddutile@redhat.com \
    --cc=helgaas@kernel.org \
    --cc=jgg@nvidia.com \
    --cc=kuba@kernel.org \
    --cc=linux-pci@vger.kernel.org \
    --cc=linux-rdma@vger.kernel.org \
    --cc=netdev@vger.kernel.org \
    --cc=saeedm@nvidia.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).