* [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
@ 2022-05-05 16:48 ` Lorenzo Bianconi
0 siblings, 0 replies; 8+ messages in thread
From: Lorenzo Bianconi @ 2022-05-05 16:48 UTC (permalink / raw)
To: netdev
Cc: intel-wired-lan, davem, kuba, pabeni, jesse.brandeburg,
anthony.l.nguyen, alice.michael, bpf, ast, daniel,
lorenzo.bianconi, andrii, magnus.karlsson, jbrouer, toke,
tirthendu.sarkar
Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
callback.
Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
---
drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
1 file changed, 62 insertions(+), 25 deletions(-)
diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
index 7bc1174edf6b..b7967105a549 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
@@ -2509,6 +2509,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget)
hard_start = page_address(rx_buffer->page) +
rx_buffer->page_offset - offset;
xdp_prepare_buff(&xdp, hard_start, offset, size, true);
+ xdp_buff_clear_frags_flag(&xdp);
#if (PAGE_SIZE > 4096)
/* At larger PAGE_SIZE, frame_sz depend on len size */
xdp.frame_sz = i40e_rx_frame_truesize(rx_ring, size);
@@ -3713,35 +3714,55 @@ u16 i40e_lan_select_queue(struct net_device *netdev,
static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
struct i40e_ring *xdp_ring)
{
- u16 i = xdp_ring->next_to_use;
- struct i40e_tx_buffer *tx_bi;
- struct i40e_tx_desc *tx_desc;
+ struct skb_shared_info *sinfo = xdp_get_shared_info_from_frame(xdpf);
+ u8 nr_frags = unlikely(xdp_frame_has_frags(xdpf)) ? sinfo->nr_frags : 0;
+ u16 i = 0, index = xdp_ring->next_to_use;
+ struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index];
+ struct i40e_tx_buffer *tx_bi = tx_head;
+ struct i40e_tx_desc *tx_desc = I40E_TX_DESC(xdp_ring, index);
void *data = xdpf->data;
u32 size = xdpf->len;
- dma_addr_t dma;
- if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) {
+ if (unlikely(I40E_DESC_UNUSED(xdp_ring) < 1 + nr_frags)) {
xdp_ring->tx_stats.tx_busy++;
return I40E_XDP_CONSUMED;
}
- dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
- if (dma_mapping_error(xdp_ring->dev, dma))
- return I40E_XDP_CONSUMED;
- tx_bi = &xdp_ring->tx_bi[i];
- tx_bi->bytecount = size;
- tx_bi->gso_segs = 1;
- tx_bi->xdpf = xdpf;
+ tx_head->bytecount = xdp_get_frame_len(xdpf);
+ tx_head->gso_segs = 1;
+ tx_head->xdpf = xdpf;
- /* record length, and DMA address */
- dma_unmap_len_set(tx_bi, len, size);
- dma_unmap_addr_set(tx_bi, dma, dma);
+ for (;;) {
+ dma_addr_t dma;
- tx_desc = I40E_TX_DESC(xdp_ring, i);
- tx_desc->buffer_addr = cpu_to_le64(dma);
- tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC
- | I40E_TXD_CMD,
- 0, size, 0);
+ dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
+ if (dma_mapping_error(xdp_ring->dev, dma))
+ goto unmap;
+
+ /* record length, and DMA address */
+ dma_unmap_len_set(tx_bi, len, size);
+ dma_unmap_addr_set(tx_bi, dma, dma);
+
+ tx_desc->buffer_addr = cpu_to_le64(dma);
+ tx_desc->cmd_type_offset_bsz =
+ build_ctob(I40E_TX_DESC_CMD_ICRC, 0, size, 0);
+
+ if (++index == xdp_ring->count)
+ index = 0;
+
+ if (i == nr_frags)
+ break;
+
+ tx_bi = &xdp_ring->tx_bi[index];
+ tx_desc = I40E_TX_DESC(xdp_ring, index);
+
+ data = skb_frag_address(&sinfo->frags[i]);
+ size = skb_frag_size(&sinfo->frags[i]);
+ i++;
+ }
+
+ tx_desc->cmd_type_offset_bsz |=
+ cpu_to_le64(I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
/* Make certain all of the status bits have been updated
* before next_to_watch is written.
@@ -3749,14 +3770,30 @@ static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
smp_wmb();
xdp_ring->xdp_tx_active++;
- i++;
- if (i == xdp_ring->count)
- i = 0;
- tx_bi->next_to_watch = tx_desc;
- xdp_ring->next_to_use = i;
+ tx_head->next_to_watch = tx_desc;
+ xdp_ring->next_to_use = index;
return I40E_XDP_TX;
+
+unmap:
+ for (;;) {
+ tx_bi = &xdp_ring->tx_bi[index];
+ if (dma_unmap_len(tx_bi, len))
+ dma_unmap_page(xdp_ring->dev,
+ dma_unmap_addr(tx_bi, dma),
+ dma_unmap_len(tx_bi, len),
+ DMA_TO_DEVICE);
+ dma_unmap_len_set(tx_bi, len, 0);
+ if (tx_bi == tx_head)
+ break;
+
+ if (!index)
+ index += xdp_ring->count;
+ index--;
+ }
+
+ return I40E_XDP_CONSUMED;
}
/**
--
2.35.1
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
@ 2022-05-05 16:48 ` Lorenzo Bianconi
0 siblings, 0 replies; 8+ messages in thread
From: Lorenzo Bianconi @ 2022-05-05 16:48 UTC (permalink / raw)
To: intel-wired-lan
Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
callback.
Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
---
drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
1 file changed, 62 insertions(+), 25 deletions(-)
diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
index 7bc1174edf6b..b7967105a549 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
@@ -2509,6 +2509,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget)
hard_start = page_address(rx_buffer->page) +
rx_buffer->page_offset - offset;
xdp_prepare_buff(&xdp, hard_start, offset, size, true);
+ xdp_buff_clear_frags_flag(&xdp);
#if (PAGE_SIZE > 4096)
/* At larger PAGE_SIZE, frame_sz depend on len size */
xdp.frame_sz = i40e_rx_frame_truesize(rx_ring, size);
@@ -3713,35 +3714,55 @@ u16 i40e_lan_select_queue(struct net_device *netdev,
static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
struct i40e_ring *xdp_ring)
{
- u16 i = xdp_ring->next_to_use;
- struct i40e_tx_buffer *tx_bi;
- struct i40e_tx_desc *tx_desc;
+ struct skb_shared_info *sinfo = xdp_get_shared_info_from_frame(xdpf);
+ u8 nr_frags = unlikely(xdp_frame_has_frags(xdpf)) ? sinfo->nr_frags : 0;
+ u16 i = 0, index = xdp_ring->next_to_use;
+ struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index];
+ struct i40e_tx_buffer *tx_bi = tx_head;
+ struct i40e_tx_desc *tx_desc = I40E_TX_DESC(xdp_ring, index);
void *data = xdpf->data;
u32 size = xdpf->len;
- dma_addr_t dma;
- if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) {
+ if (unlikely(I40E_DESC_UNUSED(xdp_ring) < 1 + nr_frags)) {
xdp_ring->tx_stats.tx_busy++;
return I40E_XDP_CONSUMED;
}
- dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
- if (dma_mapping_error(xdp_ring->dev, dma))
- return I40E_XDP_CONSUMED;
- tx_bi = &xdp_ring->tx_bi[i];
- tx_bi->bytecount = size;
- tx_bi->gso_segs = 1;
- tx_bi->xdpf = xdpf;
+ tx_head->bytecount = xdp_get_frame_len(xdpf);
+ tx_head->gso_segs = 1;
+ tx_head->xdpf = xdpf;
- /* record length, and DMA address */
- dma_unmap_len_set(tx_bi, len, size);
- dma_unmap_addr_set(tx_bi, dma, dma);
+ for (;;) {
+ dma_addr_t dma;
- tx_desc = I40E_TX_DESC(xdp_ring, i);
- tx_desc->buffer_addr = cpu_to_le64(dma);
- tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC
- | I40E_TXD_CMD,
- 0, size, 0);
+ dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
+ if (dma_mapping_error(xdp_ring->dev, dma))
+ goto unmap;
+
+ /* record length, and DMA address */
+ dma_unmap_len_set(tx_bi, len, size);
+ dma_unmap_addr_set(tx_bi, dma, dma);
+
+ tx_desc->buffer_addr = cpu_to_le64(dma);
+ tx_desc->cmd_type_offset_bsz =
+ build_ctob(I40E_TX_DESC_CMD_ICRC, 0, size, 0);
+
+ if (++index == xdp_ring->count)
+ index = 0;
+
+ if (i == nr_frags)
+ break;
+
+ tx_bi = &xdp_ring->tx_bi[index];
+ tx_desc = I40E_TX_DESC(xdp_ring, index);
+
+ data = skb_frag_address(&sinfo->frags[i]);
+ size = skb_frag_size(&sinfo->frags[i]);
+ i++;
+ }
+
+ tx_desc->cmd_type_offset_bsz |=
+ cpu_to_le64(I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
/* Make certain all of the status bits have been updated
* before next_to_watch is written.
@@ -3749,14 +3770,30 @@ static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
smp_wmb();
xdp_ring->xdp_tx_active++;
- i++;
- if (i == xdp_ring->count)
- i = 0;
- tx_bi->next_to_watch = tx_desc;
- xdp_ring->next_to_use = i;
+ tx_head->next_to_watch = tx_desc;
+ xdp_ring->next_to_use = index;
return I40E_XDP_TX;
+
+unmap:
+ for (;;) {
+ tx_bi = &xdp_ring->tx_bi[index];
+ if (dma_unmap_len(tx_bi, len))
+ dma_unmap_page(xdp_ring->dev,
+ dma_unmap_addr(tx_bi, dma),
+ dma_unmap_len(tx_bi, len),
+ DMA_TO_DEVICE);
+ dma_unmap_len_set(tx_bi, len, 0);
+ if (tx_bi == tx_head)
+ break;
+
+ if (!index)
+ index += xdp_ring->count;
+ index--;
+ }
+
+ return I40E_XDP_CONSUMED;
}
/**
--
2.35.1
^ permalink raw reply related [flat|nested] 8+ messages in thread
* Re: [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
2022-05-05 16:48 ` [Intel-wired-lan] " Lorenzo Bianconi
@ 2022-05-05 18:06 ` Paul Menzel
-1 siblings, 0 replies; 8+ messages in thread
From: Paul Menzel @ 2022-05-05 18:06 UTC (permalink / raw)
To: Lorenzo Bianconi
Cc: netdev, tirthendu.sarkar, daniel, intel-wired-lan, toke, ast,
andrii, jbrouer, kuba, bpf, pabeni, davem, magnus.karlsson
Am 05.05.22 um 18:48 schrieb Lorenzo Bianconi:
> Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
> callback.
>
> Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
> Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
> ---
> drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
> 1 file changed, 62 insertions(+), 25 deletions(-)
>
> diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> index 7bc1174edf6b..b7967105a549 100644
> --- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> +++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> @@ -2509,6 +2509,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget)
> hard_start = page_address(rx_buffer->page) +
> rx_buffer->page_offset - offset;
> xdp_prepare_buff(&xdp, hard_start, offset, size, true);
> + xdp_buff_clear_frags_flag(&xdp);
> #if (PAGE_SIZE > 4096)
> /* At larger PAGE_SIZE, frame_sz depend on len size */
> xdp.frame_sz = i40e_rx_frame_truesize(rx_ring, size);
> @@ -3713,35 +3714,55 @@ u16 i40e_lan_select_queue(struct net_device *netdev,
> static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> struct i40e_ring *xdp_ring)
> {
> - u16 i = xdp_ring->next_to_use;
> - struct i40e_tx_buffer *tx_bi;
> - struct i40e_tx_desc *tx_desc;
> + struct skb_shared_info *sinfo = xdp_get_shared_info_from_frame(xdpf);
> + u8 nr_frags = unlikely(xdp_frame_has_frags(xdpf)) ? sinfo->nr_frags : 0;
> + u16 i = 0, index = xdp_ring->next_to_use;
> + struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index];
> + struct i40e_tx_buffer *tx_bi = tx_head;
> + struct i40e_tx_desc *tx_desc = I40E_TX_DESC(xdp_ring, index);
> void *data = xdpf->data;
> u32 size = xdpf->len;
> - dma_addr_t dma;
>
> - if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) {
> + if (unlikely(I40E_DESC_UNUSED(xdp_ring) < 1 + nr_frags)) {
> xdp_ring->tx_stats.tx_busy++;
> return I40E_XDP_CONSUMED;
> }
> - dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> - if (dma_mapping_error(xdp_ring->dev, dma))
> - return I40E_XDP_CONSUMED;
>
> - tx_bi = &xdp_ring->tx_bi[i];
> - tx_bi->bytecount = size;
> - tx_bi->gso_segs = 1;
> - tx_bi->xdpf = xdpf;
> + tx_head->bytecount = xdp_get_frame_len(xdpf);
> + tx_head->gso_segs = 1;
> + tx_head->xdpf = xdpf;
>
> - /* record length, and DMA address */
> - dma_unmap_len_set(tx_bi, len, size);
> - dma_unmap_addr_set(tx_bi, dma, dma);
> + for (;;) {
> + dma_addr_t dma;
>
> - tx_desc = I40E_TX_DESC(xdp_ring, i);
> - tx_desc->buffer_addr = cpu_to_le64(dma);
> - tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC
> - | I40E_TXD_CMD,
> - 0, size, 0);
> + dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> + if (dma_mapping_error(xdp_ring->dev, dma))
> + goto unmap;
> +
> + /* record length, and DMA address */
> + dma_unmap_len_set(tx_bi, len, size);
> + dma_unmap_addr_set(tx_bi, dma, dma);
> +
> + tx_desc->buffer_addr = cpu_to_le64(dma);
> + tx_desc->cmd_type_offset_bsz =
> + build_ctob(I40E_TX_DESC_CMD_ICRC, 0, size, 0);
> +
> + if (++index == xdp_ring->count)
> + index = 0;
> +
> + if (i == nr_frags)
> + break;
> +
> + tx_bi = &xdp_ring->tx_bi[index];
> + tx_desc = I40E_TX_DESC(xdp_ring, index);
> +
> + data = skb_frag_address(&sinfo->frags[i]);
> + size = skb_frag_size(&sinfo->frags[i]);
> + i++;
> + }
> +
> + tx_desc->cmd_type_offset_bsz |=
> + cpu_to_le64(I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
>
> /* Make certain all of the status bits have been updated
> * before next_to_watch is written.
> @@ -3749,14 +3770,30 @@ static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> smp_wmb();
>
> xdp_ring->xdp_tx_active++;
> - i++;
> - if (i == xdp_ring->count)
> - i = 0;
>
> - tx_bi->next_to_watch = tx_desc;
> - xdp_ring->next_to_use = i;
> + tx_head->next_to_watch = tx_desc;
> + xdp_ring->next_to_use = index;
>
> return I40E_XDP_TX;
> +
> +unmap:
> + for (;;) {
> + tx_bi = &xdp_ring->tx_bi[index];
> + if (dma_unmap_len(tx_bi, len))
> + dma_unmap_page(xdp_ring->dev,
> + dma_unmap_addr(tx_bi, dma),
> + dma_unmap_len(tx_bi, len),
> + DMA_TO_DEVICE);
> + dma_unmap_len_set(tx_bi, len, 0);
> + if (tx_bi == tx_head)
> + break;
> +
> + if (!index)
> + index += xdp_ring->count;
> + index--;
> + }
Could
```
do {
tx_bi = &xdp_ring->tx_bi[index];
if (dma_unmap_len(tx_bi, len))
dma_unmap_page(xdp_ring->dev,
dma_unmap_addr(tx_bi, dma),
dma_unmap_len(tx_bi, len),
DMA_TO_DEVICE);
dma_unmap_len_set(tx_bi, len, 0);
if (!index)
index += xdp_ring->count;
index--;
} while (tx_bi != tx_head);
```
be used instead?
> +
> + return I40E_XDP_CONSUMED;
> }
>
> /**
Kind regards,
Paul
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
@ 2022-05-05 18:06 ` Paul Menzel
0 siblings, 0 replies; 8+ messages in thread
From: Paul Menzel @ 2022-05-05 18:06 UTC (permalink / raw)
To: intel-wired-lan
Am 05.05.22 um 18:48 schrieb Lorenzo Bianconi:
> Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
> callback.
>
> Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
> Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
> ---
> drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
> 1 file changed, 62 insertions(+), 25 deletions(-)
>
> diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> index 7bc1174edf6b..b7967105a549 100644
> --- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> +++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> @@ -2509,6 +2509,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget)
> hard_start = page_address(rx_buffer->page) +
> rx_buffer->page_offset - offset;
> xdp_prepare_buff(&xdp, hard_start, offset, size, true);
> + xdp_buff_clear_frags_flag(&xdp);
> #if (PAGE_SIZE > 4096)
> /* At larger PAGE_SIZE, frame_sz depend on len size */
> xdp.frame_sz = i40e_rx_frame_truesize(rx_ring, size);
> @@ -3713,35 +3714,55 @@ u16 i40e_lan_select_queue(struct net_device *netdev,
> static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> struct i40e_ring *xdp_ring)
> {
> - u16 i = xdp_ring->next_to_use;
> - struct i40e_tx_buffer *tx_bi;
> - struct i40e_tx_desc *tx_desc;
> + struct skb_shared_info *sinfo = xdp_get_shared_info_from_frame(xdpf);
> + u8 nr_frags = unlikely(xdp_frame_has_frags(xdpf)) ? sinfo->nr_frags : 0;
> + u16 i = 0, index = xdp_ring->next_to_use;
> + struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index];
> + struct i40e_tx_buffer *tx_bi = tx_head;
> + struct i40e_tx_desc *tx_desc = I40E_TX_DESC(xdp_ring, index);
> void *data = xdpf->data;
> u32 size = xdpf->len;
> - dma_addr_t dma;
>
> - if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) {
> + if (unlikely(I40E_DESC_UNUSED(xdp_ring) < 1 + nr_frags)) {
> xdp_ring->tx_stats.tx_busy++;
> return I40E_XDP_CONSUMED;
> }
> - dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> - if (dma_mapping_error(xdp_ring->dev, dma))
> - return I40E_XDP_CONSUMED;
>
> - tx_bi = &xdp_ring->tx_bi[i];
> - tx_bi->bytecount = size;
> - tx_bi->gso_segs = 1;
> - tx_bi->xdpf = xdpf;
> + tx_head->bytecount = xdp_get_frame_len(xdpf);
> + tx_head->gso_segs = 1;
> + tx_head->xdpf = xdpf;
>
> - /* record length, and DMA address */
> - dma_unmap_len_set(tx_bi, len, size);
> - dma_unmap_addr_set(tx_bi, dma, dma);
> + for (;;) {
> + dma_addr_t dma;
>
> - tx_desc = I40E_TX_DESC(xdp_ring, i);
> - tx_desc->buffer_addr = cpu_to_le64(dma);
> - tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC
> - | I40E_TXD_CMD,
> - 0, size, 0);
> + dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> + if (dma_mapping_error(xdp_ring->dev, dma))
> + goto unmap;
> +
> + /* record length, and DMA address */
> + dma_unmap_len_set(tx_bi, len, size);
> + dma_unmap_addr_set(tx_bi, dma, dma);
> +
> + tx_desc->buffer_addr = cpu_to_le64(dma);
> + tx_desc->cmd_type_offset_bsz =
> + build_ctob(I40E_TX_DESC_CMD_ICRC, 0, size, 0);
> +
> + if (++index == xdp_ring->count)
> + index = 0;
> +
> + if (i == nr_frags)
> + break;
> +
> + tx_bi = &xdp_ring->tx_bi[index];
> + tx_desc = I40E_TX_DESC(xdp_ring, index);
> +
> + data = skb_frag_address(&sinfo->frags[i]);
> + size = skb_frag_size(&sinfo->frags[i]);
> + i++;
> + }
> +
> + tx_desc->cmd_type_offset_bsz |=
> + cpu_to_le64(I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
>
> /* Make certain all of the status bits have been updated
> * before next_to_watch is written.
> @@ -3749,14 +3770,30 @@ static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> smp_wmb();
>
> xdp_ring->xdp_tx_active++;
> - i++;
> - if (i == xdp_ring->count)
> - i = 0;
>
> - tx_bi->next_to_watch = tx_desc;
> - xdp_ring->next_to_use = i;
> + tx_head->next_to_watch = tx_desc;
> + xdp_ring->next_to_use = index;
>
> return I40E_XDP_TX;
> +
> +unmap:
> + for (;;) {
> + tx_bi = &xdp_ring->tx_bi[index];
> + if (dma_unmap_len(tx_bi, len))
> + dma_unmap_page(xdp_ring->dev,
> + dma_unmap_addr(tx_bi, dma),
> + dma_unmap_len(tx_bi, len),
> + DMA_TO_DEVICE);
> + dma_unmap_len_set(tx_bi, len, 0);
> + if (tx_bi == tx_head)
> + break;
> +
> + if (!index)
> + index += xdp_ring->count;
> + index--;
> + }
Could
```
do {
tx_bi = &xdp_ring->tx_bi[index];
if (dma_unmap_len(tx_bi, len))
dma_unmap_page(xdp_ring->dev,
dma_unmap_addr(tx_bi, dma),
dma_unmap_len(tx_bi, len),
DMA_TO_DEVICE);
dma_unmap_len_set(tx_bi, len, 0);
if (!index)
index += xdp_ring->count;
index--;
} while (tx_bi != tx_head);
```
be used instead?
> +
> + return I40E_XDP_CONSUMED;
> }
>
> /**
Kind regards,
Paul
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
2022-05-05 18:06 ` Paul Menzel
@ 2022-05-05 20:09 ` Lorenzo Bianconi
-1 siblings, 0 replies; 8+ messages in thread
From: Lorenzo Bianconi @ 2022-05-05 20:09 UTC (permalink / raw)
To: Paul Menzel
Cc: Lorenzo Bianconi, netdev, tirthendu.sarkar, daniel,
intel-wired-lan, toke, ast, andrii, jbrouer, kuba, bpf, pabeni,
davem, magnus.karlsson
[-- Attachment #1: Type: text/plain, Size: 5480 bytes --]
>
>
> Am 05.05.22 um 18:48 schrieb Lorenzo Bianconi:
> > Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
> > callback.
> >
> > Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
> > Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
> > ---
> > drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
> > 1 file changed, 62 insertions(+), 25 deletions(-)
> >
> > diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> > index 7bc1174edf6b..b7967105a549 100644
> > --- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> > +++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> > @@ -2509,6 +2509,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget)
> > hard_start = page_address(rx_buffer->page) +
> > rx_buffer->page_offset - offset;
> > xdp_prepare_buff(&xdp, hard_start, offset, size, true);
> > + xdp_buff_clear_frags_flag(&xdp);
> > #if (PAGE_SIZE > 4096)
> > /* At larger PAGE_SIZE, frame_sz depend on len size */
> > xdp.frame_sz = i40e_rx_frame_truesize(rx_ring, size);
> > @@ -3713,35 +3714,55 @@ u16 i40e_lan_select_queue(struct net_device *netdev,
> > static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> > struct i40e_ring *xdp_ring)
> > {
> > - u16 i = xdp_ring->next_to_use;
> > - struct i40e_tx_buffer *tx_bi;
> > - struct i40e_tx_desc *tx_desc;
> > + struct skb_shared_info *sinfo = xdp_get_shared_info_from_frame(xdpf);
> > + u8 nr_frags = unlikely(xdp_frame_has_frags(xdpf)) ? sinfo->nr_frags : 0;
> > + u16 i = 0, index = xdp_ring->next_to_use;
> > + struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index];
> > + struct i40e_tx_buffer *tx_bi = tx_head;
> > + struct i40e_tx_desc *tx_desc = I40E_TX_DESC(xdp_ring, index);
> > void *data = xdpf->data;
> > u32 size = xdpf->len;
> > - dma_addr_t dma;
> > - if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) {
> > + if (unlikely(I40E_DESC_UNUSED(xdp_ring) < 1 + nr_frags)) {
> > xdp_ring->tx_stats.tx_busy++;
> > return I40E_XDP_CONSUMED;
> > }
> > - dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> > - if (dma_mapping_error(xdp_ring->dev, dma))
> > - return I40E_XDP_CONSUMED;
> > - tx_bi = &xdp_ring->tx_bi[i];
> > - tx_bi->bytecount = size;
> > - tx_bi->gso_segs = 1;
> > - tx_bi->xdpf = xdpf;
> > + tx_head->bytecount = xdp_get_frame_len(xdpf);
> > + tx_head->gso_segs = 1;
> > + tx_head->xdpf = xdpf;
> > - /* record length, and DMA address */
> > - dma_unmap_len_set(tx_bi, len, size);
> > - dma_unmap_addr_set(tx_bi, dma, dma);
> > + for (;;) {
> > + dma_addr_t dma;
> > - tx_desc = I40E_TX_DESC(xdp_ring, i);
> > - tx_desc->buffer_addr = cpu_to_le64(dma);
> > - tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC
> > - | I40E_TXD_CMD,
> > - 0, size, 0);
> > + dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> > + if (dma_mapping_error(xdp_ring->dev, dma))
> > + goto unmap;
> > +
> > + /* record length, and DMA address */
> > + dma_unmap_len_set(tx_bi, len, size);
> > + dma_unmap_addr_set(tx_bi, dma, dma);
> > +
> > + tx_desc->buffer_addr = cpu_to_le64(dma);
> > + tx_desc->cmd_type_offset_bsz =
> > + build_ctob(I40E_TX_DESC_CMD_ICRC, 0, size, 0);
> > +
> > + if (++index == xdp_ring->count)
> > + index = 0;
> > +
> > + if (i == nr_frags)
> > + break;
> > +
> > + tx_bi = &xdp_ring->tx_bi[index];
> > + tx_desc = I40E_TX_DESC(xdp_ring, index);
> > +
> > + data = skb_frag_address(&sinfo->frags[i]);
> > + size = skb_frag_size(&sinfo->frags[i]);
> > + i++;
> > + }
> > +
> > + tx_desc->cmd_type_offset_bsz |=
> > + cpu_to_le64(I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
> > /* Make certain all of the status bits have been updated
> > * before next_to_watch is written.
> > @@ -3749,14 +3770,30 @@ static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> > smp_wmb();
> > xdp_ring->xdp_tx_active++;
> > - i++;
> > - if (i == xdp_ring->count)
> > - i = 0;
> > - tx_bi->next_to_watch = tx_desc;
> > - xdp_ring->next_to_use = i;
> > + tx_head->next_to_watch = tx_desc;
> > + xdp_ring->next_to_use = index;
> > return I40E_XDP_TX;
> > +
> > +unmap:
> > + for (;;) {
> > + tx_bi = &xdp_ring->tx_bi[index];
> > + if (dma_unmap_len(tx_bi, len))
> > + dma_unmap_page(xdp_ring->dev,
> > + dma_unmap_addr(tx_bi, dma),
> > + dma_unmap_len(tx_bi, len),
> > + DMA_TO_DEVICE);
> > + dma_unmap_len_set(tx_bi, len, 0);
> > + if (tx_bi == tx_head)
> > + break;
> > +
> > + if (!index)
> > + index += xdp_ring->count;
> > + index--;
> > + }
>
> Could
>
> ```
> do {
> tx_bi = &xdp_ring->tx_bi[index];
> if (dma_unmap_len(tx_bi, len))
> dma_unmap_page(xdp_ring->dev,
> dma_unmap_addr(tx_bi, dma),
> dma_unmap_len(tx_bi, len),
> DMA_TO_DEVICE);
> dma_unmap_len_set(tx_bi, len, 0);
>
> if (!index)
> index += xdp_ring->count;
> index--;
> } while (tx_bi != tx_head);
> ```
>
> be used instead?
yes, it seems just a matter of test to me, doesn't it? :)
Regards,
Lorenzo
>
> > +
> > + return I40E_XDP_CONSUMED;
> > }
> > /**
>
>
> Kind regards,
>
> Paul
>
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 228 bytes --]
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
@ 2022-05-05 20:09 ` Lorenzo Bianconi
0 siblings, 0 replies; 8+ messages in thread
From: Lorenzo Bianconi @ 2022-05-05 20:09 UTC (permalink / raw)
To: intel-wired-lan
>
>
> Am 05.05.22 um 18:48 schrieb Lorenzo Bianconi:
> > Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
> > callback.
> >
> > Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
> > Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
> > ---
> > drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
> > 1 file changed, 62 insertions(+), 25 deletions(-)
> >
> > diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> > index 7bc1174edf6b..b7967105a549 100644
> > --- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> > +++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
> > @@ -2509,6 +2509,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget)
> > hard_start = page_address(rx_buffer->page) +
> > rx_buffer->page_offset - offset;
> > xdp_prepare_buff(&xdp, hard_start, offset, size, true);
> > + xdp_buff_clear_frags_flag(&xdp);
> > #if (PAGE_SIZE > 4096)
> > /* At larger PAGE_SIZE, frame_sz depend on len size */
> > xdp.frame_sz = i40e_rx_frame_truesize(rx_ring, size);
> > @@ -3713,35 +3714,55 @@ u16 i40e_lan_select_queue(struct net_device *netdev,
> > static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> > struct i40e_ring *xdp_ring)
> > {
> > - u16 i = xdp_ring->next_to_use;
> > - struct i40e_tx_buffer *tx_bi;
> > - struct i40e_tx_desc *tx_desc;
> > + struct skb_shared_info *sinfo = xdp_get_shared_info_from_frame(xdpf);
> > + u8 nr_frags = unlikely(xdp_frame_has_frags(xdpf)) ? sinfo->nr_frags : 0;
> > + u16 i = 0, index = xdp_ring->next_to_use;
> > + struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index];
> > + struct i40e_tx_buffer *tx_bi = tx_head;
> > + struct i40e_tx_desc *tx_desc = I40E_TX_DESC(xdp_ring, index);
> > void *data = xdpf->data;
> > u32 size = xdpf->len;
> > - dma_addr_t dma;
> > - if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) {
> > + if (unlikely(I40E_DESC_UNUSED(xdp_ring) < 1 + nr_frags)) {
> > xdp_ring->tx_stats.tx_busy++;
> > return I40E_XDP_CONSUMED;
> > }
> > - dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> > - if (dma_mapping_error(xdp_ring->dev, dma))
> > - return I40E_XDP_CONSUMED;
> > - tx_bi = &xdp_ring->tx_bi[i];
> > - tx_bi->bytecount = size;
> > - tx_bi->gso_segs = 1;
> > - tx_bi->xdpf = xdpf;
> > + tx_head->bytecount = xdp_get_frame_len(xdpf);
> > + tx_head->gso_segs = 1;
> > + tx_head->xdpf = xdpf;
> > - /* record length, and DMA address */
> > - dma_unmap_len_set(tx_bi, len, size);
> > - dma_unmap_addr_set(tx_bi, dma, dma);
> > + for (;;) {
> > + dma_addr_t dma;
> > - tx_desc = I40E_TX_DESC(xdp_ring, i);
> > - tx_desc->buffer_addr = cpu_to_le64(dma);
> > - tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC
> > - | I40E_TXD_CMD,
> > - 0, size, 0);
> > + dma = dma_map_single(xdp_ring->dev, data, size, DMA_TO_DEVICE);
> > + if (dma_mapping_error(xdp_ring->dev, dma))
> > + goto unmap;
> > +
> > + /* record length, and DMA address */
> > + dma_unmap_len_set(tx_bi, len, size);
> > + dma_unmap_addr_set(tx_bi, dma, dma);
> > +
> > + tx_desc->buffer_addr = cpu_to_le64(dma);
> > + tx_desc->cmd_type_offset_bsz =
> > + build_ctob(I40E_TX_DESC_CMD_ICRC, 0, size, 0);
> > +
> > + if (++index == xdp_ring->count)
> > + index = 0;
> > +
> > + if (i == nr_frags)
> > + break;
> > +
> > + tx_bi = &xdp_ring->tx_bi[index];
> > + tx_desc = I40E_TX_DESC(xdp_ring, index);
> > +
> > + data = skb_frag_address(&sinfo->frags[i]);
> > + size = skb_frag_size(&sinfo->frags[i]);
> > + i++;
> > + }
> > +
> > + tx_desc->cmd_type_offset_bsz |=
> > + cpu_to_le64(I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
> > /* Make certain all of the status bits have been updated
> > * before next_to_watch is written.
> > @@ -3749,14 +3770,30 @@ static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
> > smp_wmb();
> > xdp_ring->xdp_tx_active++;
> > - i++;
> > - if (i == xdp_ring->count)
> > - i = 0;
> > - tx_bi->next_to_watch = tx_desc;
> > - xdp_ring->next_to_use = i;
> > + tx_head->next_to_watch = tx_desc;
> > + xdp_ring->next_to_use = index;
> > return I40E_XDP_TX;
> > +
> > +unmap:
> > + for (;;) {
> > + tx_bi = &xdp_ring->tx_bi[index];
> > + if (dma_unmap_len(tx_bi, len))
> > + dma_unmap_page(xdp_ring->dev,
> > + dma_unmap_addr(tx_bi, dma),
> > + dma_unmap_len(tx_bi, len),
> > + DMA_TO_DEVICE);
> > + dma_unmap_len_set(tx_bi, len, 0);
> > + if (tx_bi == tx_head)
> > + break;
> > +
> > + if (!index)
> > + index += xdp_ring->count;
> > + index--;
> > + }
>
> Could
>
> ```
> do {
> tx_bi = &xdp_ring->tx_bi[index];
> if (dma_unmap_len(tx_bi, len))
> dma_unmap_page(xdp_ring->dev,
> dma_unmap_addr(tx_bi, dma),
> dma_unmap_len(tx_bi, len),
> DMA_TO_DEVICE);
> dma_unmap_len_set(tx_bi, len, 0);
>
> if (!index)
> index += xdp_ring->count;
> index--;
> } while (tx_bi != tx_head);
> ```
>
> be used instead?
yes, it seems just a matter of test to me, doesn't it? :)
Regards,
Lorenzo
>
> > +
> > + return I40E_XDP_CONSUMED;
> > }
> > /**
>
>
> Kind regards,
>
> Paul
>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 228 bytes
Desc: not available
URL: <http://lists.osuosl.org/pipermail/intel-wired-lan/attachments/20220505/5a4e9ed2/attachment-0001.asc>
^ permalink raw reply [flat|nested] 8+ messages in thread
* RE: [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
2022-05-05 16:48 ` [Intel-wired-lan] " Lorenzo Bianconi
@ 2022-06-03 7:14 ` Kuruvinakunnel, George
-1 siblings, 0 replies; 8+ messages in thread
From: Kuruvinakunnel, George @ 2022-06-03 7:14 UTC (permalink / raw)
To: Lorenzo Bianconi, netdev
Cc: Sarkar, Tirthendu, daniel, intel-wired-lan, toke, ast, andrii,
jbrouer, kuba, bpf, pabeni, davem, Karlsson, Magnus
> -----Original Message-----
> From: Intel-wired-lan <intel-wired-lan-bounces@osuosl.org> On Behalf Of Lorenzo
> Bianconi
> Sent: Thursday, May 5, 2022 10:19 PM
> To: netdev@vger.kernel.org
> Cc: Sarkar, Tirthendu <tirthendu.sarkar@intel.com>; daniel@iogearbox.net; intel-
> wired-lan@lists.osuosl.org; toke@redhat.com; ast@kernel.org; andrii@kernel.org;
> jbrouer@redhat.com; kuba@kernel.org; bpf@vger.kernel.org; pabeni@redhat.com;
> davem@davemloft.net; Karlsson, Magnus <magnus.karlsson@intel.com>
> Subject: [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to
> ndo_xdp_xmit
>
> Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
> callback.
>
> Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
> Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
> ---
> drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
> 1 file changed, 62 insertions(+), 25 deletions(-)
>
Tested-by: George Kuruvinakunnel <george.kuruvinakunnel@intel.com>
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit
@ 2022-06-03 7:14 ` Kuruvinakunnel, George
0 siblings, 0 replies; 8+ messages in thread
From: Kuruvinakunnel, George @ 2022-06-03 7:14 UTC (permalink / raw)
To: Lorenzo Bianconi, netdev
Cc: Sarkar, Tirthendu, daniel, toke, andrii, jbrouer, ast,
intel-wired-lan, kuba, bpf, pabeni, davem, Karlsson, Magnus
> -----Original Message-----
> From: Intel-wired-lan <intel-wired-lan-bounces@osuosl.org> On Behalf Of Lorenzo
> Bianconi
> Sent: Thursday, May 5, 2022 10:19 PM
> To: netdev@vger.kernel.org
> Cc: Sarkar, Tirthendu <tirthendu.sarkar@intel.com>; daniel@iogearbox.net; intel-
> wired-lan@lists.osuosl.org; toke@redhat.com; ast@kernel.org; andrii@kernel.org;
> jbrouer@redhat.com; kuba@kernel.org; bpf@vger.kernel.org; pabeni@redhat.com;
> davem@davemloft.net; Karlsson, Magnus <magnus.karlsson@intel.com>
> Subject: [Intel-wired-lan] [PATCH net-next] i40e: add xdp frags support to
> ndo_xdp_xmit
>
> Add the capability to map non-linear xdp frames in XDP_TX and ndo_xdp_xmit
> callback.
>
> Tested-by: Sarkar Tirthendu <tirthendu.sarkar@intel.com>
> Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
> ---
> drivers/net/ethernet/intel/i40e/i40e_txrx.c | 87 +++++++++++++++------
> 1 file changed, 62 insertions(+), 25 deletions(-)
>
Tested-by: George Kuruvinakunnel <george.kuruvinakunnel@intel.com>
_______________________________________________
Intel-wired-lan mailing list
Intel-wired-lan@osuosl.org
https://lists.osuosl.org/mailman/listinfo/intel-wired-lan
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2022-06-03 7:14 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-05-05 16:48 [PATCH net-next] i40e: add xdp frags support to ndo_xdp_xmit Lorenzo Bianconi
2022-05-05 16:48 ` [Intel-wired-lan] " Lorenzo Bianconi
2022-05-05 18:06 ` Paul Menzel
2022-05-05 18:06 ` Paul Menzel
2022-05-05 20:09 ` Lorenzo Bianconi
2022-05-05 20:09 ` Lorenzo Bianconi
2022-06-03 7:14 ` Kuruvinakunnel, George
2022-06-03 7:14 ` Kuruvinakunnel, George
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.