bpf.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Heng Qi <hengqi@linux.alibaba.com>
To: Gavin Li <gavinl@nvidia.com>
Cc: gavi@nvidia.com, virtualization@lists.linux-foundation.org,
	netdev@vger.kernel.org, linux-kernel@vger.kernel.org,
	bpf@vger.kernel.org, mst@redhat.com, jasowang@redhat.com,
	xuanzhuo@linux.alibaba.com, davem@davemloft.net,
	edumazet@google.com, kuba@kernel.org, pabeni@redhat.com,
	ast@kernel.org, daniel@iogearbox.net, hawk@kernel.org,
	john.fastabend@gmail.com, jiri@nvidia.com, dtatulea@nvidia.com
Subject: Re: [PATCH net-next V2 3/4] virtio_net: support per queue interrupt coalesce command
Date: Wed, 19 Jul 2023 20:19:27 +0800	[thread overview]
Message-ID: <291a6897-d706-688d-ca4a-832ee12eb402@linux.alibaba.com> (raw)
In-Reply-To: <a2cd512a-245e-4c8a-633d-126c6fc135ba@nvidia.com>



在 2023/7/18 下午2:45, Gavin Li 写道:
>
>
> On 7/18/2023 11:37 AM, Heng Qi wrote:
>>
>>
>> 在 2023/7/17 下午10:30, Gavin Li 写道:
>>> Add interrupt_coalesce config in send_queue and receive_queue to 
>>> cache user
>>> config.
>>>
>>> Send per virtqueue interrupt moderation config to underline device 
>>> in order
>>> to have more efficient interrupt moderation and cpu utilization of 
>>> guest
>>> VM.
>>>
>>> Signed-off-by: Gavin Li <gavinl@nvidia.com>
>>> Reviewed-by: Dragos Tatulea <dtatulea@nvidia.com>
>>> Reviewed-by: Jiri Pirko <jiri@nvidia.com>
>>> Acked-by: Michael S. Tsirkin <mst@redhat.com>
>>> ---
>>>   drivers/net/virtio_net.c        | 123 
>>> ++++++++++++++++++++++++++++----
>>>   include/uapi/linux/virtio_net.h |  14 ++++
>>>   2 files changed, 125 insertions(+), 12 deletions(-)
>>>
>>> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
>>> index 802ed21453f5..1566c7de9436 100644
>>> --- a/drivers/net/virtio_net.c
>>> +++ b/drivers/net/virtio_net.c
>>> @@ -144,6 +144,8 @@ struct send_queue {
>>>       struct virtnet_sq_stats stats;
>>> +    struct virtnet_interrupt_coalesce intr_coal;
>>> +
>>>       struct napi_struct napi;
>>>       /* Record whether sq is in reset state. */
>>> @@ -161,6 +163,8 @@ struct receive_queue {
>>>       struct virtnet_rq_stats stats;
>>> +    struct virtnet_interrupt_coalesce intr_coal;
>>> +
>>>       /* Chain pages by the private ptr. */
>>>       struct page *pages;
>>> @@ -3078,6 +3082,59 @@ static int virtnet_send_notf_coal_cmds(struct 
>>> virtnet_info *vi,
>>>       return 0;
>>>   }
>>> +static int virtnet_send_ctrl_coal_vq_cmd(struct virtnet_info *vi,
>>> +                     u16 vqn, u32 max_usecs, u32 max_packets)
>>> +{
>>> +    struct virtio_net_ctrl_coal_vq *coal_vq;
>>> +    struct scatterlist sgs;
>>> +
>>> +    coal_vq = kzalloc(sizeof(*coal_vq), GFP_KERNEL);
>>> +    if (!coal_vq)
>>> +        return -ENOMEM;
>>> +    coal_vq->vqn = cpu_to_le16(vqn);
>>> +    coal_vq->coal.max_usecs = cpu_to_le32(max_usecs);
>>> +    coal_vq->coal.max_packets = cpu_to_le32(max_packets);
>>> +    sg_init_one(&sgs, coal_vq, sizeof(*coal_vq));
>>> +
>>> +    if (!virtnet_send_command(vi, VIRTIO_NET_CTRL_NOTF_COAL,
>>> +                  VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET,
>>> +                  &sgs))
>>> +        return -EINVAL;
>>> +
>>> +    return 0;
>>> +}
>>> +
>>> +static int virtnet_send_notf_coal_vq_cmds(struct virtnet_info *vi,
>>> +                      struct ethtool_coalesce *ec,
>>> +                      u16 queue)
>>> +{
>>> +    int err;
>>> +
>>> +    if (ec->rx_coalesce_usecs || ec->rx_max_coalesced_frames) {
>>> +        err = virtnet_send_ctrl_coal_vq_cmd(vi, rxq2vq(queue),
>>> +                            ec->rx_coalesce_usecs,
>>> +                            ec->rx_max_coalesced_frames);
>>> +        if (err)
>>> +            return err;
>>> +        /* Save parameters */
>>> +        vi->rq[queue].intr_coal.max_usecs = ec->rx_coalesce_usecs;
>>> +        vi->rq[queue].intr_coal.max_packets = 
>>> ec->rx_max_coalesced_frames;
>>> +    }
>>> +
>>> +    if (ec->tx_coalesce_usecs || ec->tx_max_coalesced_frames) {
>>> +        err = virtnet_send_ctrl_coal_vq_cmd(vi, txq2vq(queue),
>>> +                            ec->tx_coalesce_usecs,
>>> +                            ec->tx_max_coalesced_frames);
>>> +        if (err)
>>> +            return err;
>>> +        /* Save parameters */
>>> +        vi->sq[queue].intr_coal.max_usecs = ec->tx_coalesce_usecs;
>>> +        vi->sq[queue].intr_coal.max_packets = 
>>> ec->tx_max_coalesced_frames;
>>> +    }
>>> +
>>> +    return 0;
>>> +}
>>> +
>>>   static int virtnet_coal_params_supported(struct ethtool_coalesce *ec)
>>>   {
>>>       /* usecs coalescing is supported only if VIRTIO_NET_F_NOTF_COAL
>>> @@ -3094,23 +3151,39 @@ static int 
>>> virtnet_coal_params_supported(struct ethtool_coalesce *ec)
>>>   }
>>>   static int virtnet_set_coalesce_one(struct net_device *dev,
>>> -                    struct ethtool_coalesce *ec)
>>> +                    struct ethtool_coalesce *ec,
>>> +                    bool per_queue,
>>> +                    u32 queue)
>>>   {
>>>       struct virtnet_info *vi = netdev_priv(dev);
>>> -    int ret, i, napi_weight;
>>> +    int queue_count = per_queue ? 1 : vi->max_queue_pairs;
>>> +    int queue_number = per_queue ? queue : 0;
>>>       bool update_napi = false;
>>> +    int ret, i, napi_weight;
>>> +
>>> +    if (queue >= vi->max_queue_pairs)
>>> +        return -EINVAL;
>>>       /* Can't change NAPI weight if the link is up */
>>>       napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0;
>>> -    if (napi_weight ^ vi->sq[0].napi.weight) {
>>> -        if (dev->flags & IFF_UP)
>>> -            return -EBUSY;
>>> -        else
>>> +    for (i = queue_number; i < queue_count; i++) {
>>> +        if (napi_weight ^ vi->sq[i].napi.weight) {
>>> +            if (dev->flags & IFF_UP)
>>> +                return -EBUSY;
>>> +
>>>               update_napi = true;
>>> +            /* All queues that belong to [queue_number, 
>>> queue_count] will be
>>> +             * updated for the sake of simplicity, which might not 
>>> be necessary
>>> +             */
>>> +            queue_number = i;
>>> +            break;
>>> +        }
>>>       }
>>> -    if (virtio_has_feature(vi->vdev, VIRTIO_NET_F_NOTF_COAL))
>>> +    if (!per_queue && virtio_has_feature(vi->vdev, 
>>> VIRTIO_NET_F_NOTF_COAL))
>>>           ret = virtnet_send_notf_coal_cmds(vi, ec);
>>> +    else if (per_queue && virtio_has_feature(vi->vdev, 
>>> VIRTIO_NET_F_VQ_NOTF_COAL))
>>> +        ret = virtnet_send_notf_coal_vq_cmds(vi, ec, queue);
>>>       else
>>>           ret = virtnet_coal_params_supported(ec);
>>> @@ -3118,7 +3191,7 @@ static int virtnet_set_coalesce_one(struct 
>>> net_device *dev,
>>>           return ret;
>>>       if (update_napi) {
>>> -        for (i = 0; i < vi->max_queue_pairs; i++)
>>> +        for (i = queue_number; i < queue_count; i++)
>>>               vi->sq[i].napi.weight = napi_weight;
>>>       }
>>> @@ -3130,19 +3203,29 @@ static int virtnet_set_coalesce(struct 
>>> net_device *dev,
>>>                   struct kernel_ethtool_coalesce *kernel_coal,
>>>                   struct netlink_ext_ack *extack)
>>>   {
>>> -    return virtnet_set_coalesce_one(dev, ec);
>>> +    return virtnet_set_coalesce_one(dev, ec, false, 0);
>>>   }
>>>   static int virtnet_get_coalesce_one(struct net_device *dev,
>>> -                    struct ethtool_coalesce *ec)
>>> +                    struct ethtool_coalesce *ec,
>>> +                    bool per_queue,
>>> +                    u32 queue)
>>>   {
>>>       struct virtnet_info *vi = netdev_priv(dev);
>>> -    if (virtio_has_feature(vi->vdev, VIRTIO_NET_F_NOTF_COAL)) {
>>> +    if (queue >= vi->max_queue_pairs)
>>> +        return -EINVAL;
>>> +
>>> +    if (!per_queue && virtio_has_feature(vi->vdev, 
>>> VIRTIO_NET_F_NOTF_COAL)) {
>>>           ec->rx_coalesce_usecs = vi->intr_coal_rx.max_usecs;
>>>           ec->tx_coalesce_usecs = vi->intr_coal_tx.max_usecs;
>>>           ec->tx_max_coalesced_frames = vi->intr_coal_tx.max_packets;
>>>           ec->rx_max_coalesced_frames = vi->intr_coal_rx.max_packets;
>>> +    } else if (per_queue && virtio_has_feature(vi->vdev, 
>>> VIRTIO_NET_F_VQ_NOTF_COAL)) {
>>> +        ec->rx_coalesce_usecs = vi->rq[queue].intr_coal.max_usecs;
>>> +        ec->tx_coalesce_usecs = vi->sq[queue].intr_coal.max_usecs;
>>> +        ec->tx_max_coalesced_frames = 
>>> vi->sq[queue].intr_coal.max_packets;
>>> +        ec->rx_max_coalesced_frames = 
>>> vi->rq[queue].intr_coal.max_packets;
>>>       } else {
>>>           ec->rx_max_coalesced_frames = 1;
>>> @@ -3158,7 +3241,21 @@ static int virtnet_get_coalesce(struct 
>>> net_device *dev,
>>>                   struct kernel_ethtool_coalesce *kernel_coal,
>>>                   struct netlink_ext_ack *extack)
>>>   {
>>> -    return virtnet_get_coalesce_one(dev, ec);
>>> +    return virtnet_get_coalesce_one(dev, ec, false, 0);
>>> +}
>>> +
>>> +static int virtnet_set_per_queue_coalesce(struct net_device *dev,
>>> +                      u32 queue,
>>> +                      struct ethtool_coalesce *ec)
>>
>> When \field{max_virtqueue_pairs} is the maximum value, and the user 
>> does not carry the queue_mask for 'ethtool -Q',
>> we will send same command for all vqs, and the device will receive a 
>> large number of the same VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET commands at 
>> this time.
>> Do we want to alleviate this situation?
>>
>> Thanks.
>>
>
> May it be better to handle this senario in ethtool user space and call 
> set_coalesce instead of set_per_queue_coalesce? I'm not sure.

Maybe we need to force 'ethtool -Q' to carry queue_mask in userspace or 
instead call set_coalesce() interface when not carrying queue_mask
(this does not hinder this work and should be in another thread, please 
move on). But this I'm not sure either, since other NICs don't seem to
care much about this. Simply check that all drivers that implement the 
set_per_queue_coalesce() interface have implemented set_coalesce().

Thanks.

>>> +{
>>> +    return virtnet_set_coalesce_one(dev, ec, true, queue);
>>> +}
>>> +
>>> +static int virtnet_get_per_queue_coalesce(struct net_device *dev,
>>> +                      u32 queue,
>>> +                      struct ethtool_coalesce *ec)
>>> +{
>>> +    return virtnet_get_coalesce_one(dev, ec, true, queue);
>>>   }
>>>   static void virtnet_init_settings(struct net_device *dev)
>>> @@ -3291,6 +3388,8 @@ static const struct ethtool_ops 
>>> virtnet_ethtool_ops = {
>>>       .set_link_ksettings = virtnet_set_link_ksettings,
>>>       .set_coalesce = virtnet_set_coalesce,
>>>       .get_coalesce = virtnet_get_coalesce,
>>> +    .set_per_queue_coalesce = virtnet_set_per_queue_coalesce,
>>> +    .get_per_queue_coalesce = virtnet_get_per_queue_coalesce,
>>>       .get_rxfh_key_size = virtnet_get_rxfh_key_size,
>>>       .get_rxfh_indir_size = virtnet_get_rxfh_indir_size,
>>>       .get_rxfh = virtnet_get_rxfh,
>>> diff --git a/include/uapi/linux/virtio_net.h 
>>> b/include/uapi/linux/virtio_net.h
>>> index 12c1c9699935..cc65ef0f3c3e 100644
>>> --- a/include/uapi/linux/virtio_net.h
>>> +++ b/include/uapi/linux/virtio_net.h
>>> @@ -56,6 +56,7 @@
>>>   #define VIRTIO_NET_F_MQ    22    /* Device supports Receive Flow
>>>                        * Steering */
>>>   #define VIRTIO_NET_F_CTRL_MAC_ADDR 23    /* Set MAC address */
>>> +#define VIRTIO_NET_F_VQ_NOTF_COAL 52    /* Device supports 
>>> virtqueue notification coalescing */
>>>   #define VIRTIO_NET_F_NOTF_COAL    53    /* Device supports 
>>> notifications coalescing */
>>>   #define VIRTIO_NET_F_GUEST_USO4    54    /* Guest can handle USOv4 
>>> in. */
>>>   #define VIRTIO_NET_F_GUEST_USO6    55    /* Guest can handle USOv6 
>>> in. */
>>> @@ -391,5 +392,18 @@ struct virtio_net_ctrl_coal_rx {
>>>   };
>>>   #define VIRTIO_NET_CTRL_NOTF_COAL_RX_SET        1
>>> +#define VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET        2
>>> +#define VIRTIO_NET_CTRL_NOTF_COAL_VQ_GET        3
>>> +
>>> +struct virtio_net_ctrl_coal {
>>> +    __le32 max_packets;
>>> +    __le32 max_usecs;
>>> +};
>>> +
>>> +struct  virtio_net_ctrl_coal_vq {
>>> +    __le16 vqn;
>>> +    __le16 reserved;
>>> +    struct virtio_net_ctrl_coal coal;
>>> +};
>>>   #endif /* _UAPI_LINUX_VIRTIO_NET_H */
>>


  reply	other threads:[~2023-07-19 12:19 UTC|newest]

Thread overview: 11+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-07-17 14:30 [PATCH net-next V2 0/4] virtio_net: add per queue interrupt coalescing support Gavin Li
2023-07-17 14:30 ` [PATCH net-next V2 1/4] virtio_net: extract interrupt coalescing settings to a structure Gavin Li
2023-07-17 14:30 ` [PATCH net-next V2 2/4] virtio_net: extract get/set interrupt coalesce to a function Gavin Li
2023-07-17 14:30 ` [PATCH net-next V2 3/4] virtio_net: support per queue interrupt coalesce command Gavin Li
2023-07-17 15:22   ` Gavin Li
2023-07-18  3:29   ` Heng Qi
2023-07-18  6:28     ` Gavin Li
2023-07-18  3:37   ` Heng Qi
2023-07-18  6:45     ` Gavin Li
2023-07-19 12:19       ` Heng Qi [this message]
2023-07-17 14:30 ` [PATCH net-next V2 4/4] virtio_net: enable per queue interrupt coalesce feature Gavin Li

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=291a6897-d706-688d-ca4a-832ee12eb402@linux.alibaba.com \
    --to=hengqi@linux.alibaba.com \
    --cc=ast@kernel.org \
    --cc=bpf@vger.kernel.org \
    --cc=daniel@iogearbox.net \
    --cc=davem@davemloft.net \
    --cc=dtatulea@nvidia.com \
    --cc=edumazet@google.com \
    --cc=gavi@nvidia.com \
    --cc=gavinl@nvidia.com \
    --cc=hawk@kernel.org \
    --cc=jasowang@redhat.com \
    --cc=jiri@nvidia.com \
    --cc=john.fastabend@gmail.com \
    --cc=kuba@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=mst@redhat.com \
    --cc=netdev@vger.kernel.org \
    --cc=pabeni@redhat.com \
    --cc=virtualization@lists.linux-foundation.org \
    --cc=xuanzhuo@linux.alibaba.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).