Re: [PATCH bpf v2 7/9] idpf: use truesize as XDP RxQ info frag_size

From: Alexander Lobakin

Date: Mon Feb 16 2026 - 10:26:20 EST


From: Zaremba, Larysa <larysa.zaremba@xxxxxxxxx>
Date: Mon, 16 Feb 2026 15:01:45 +0100

> On Mon, Feb 16, 2026 at 11:48:27AM +0100, Alexander Lobakin wrote:
>> From: Alexander Lobakin <aleksander.lobakin@xxxxxxxxx>
>> Date: Mon, 16 Feb 2026 11:46:05 +0100
>>
>>> From: Zaremba, Larysa <larysa.zaremba@xxxxxxxxx>
>>> Date: Thu, 12 Feb 2026 19:33:22 +0100
>>>
>>>> The only user of frag_size field in XDP RxQ info is
>>>> bpf_xdp_frags_increase_tail(). It clearly expects whole buffer size instead
>>>> of DMA write size. Different assumptions in idpf driver configuration lead
>>>> to negative tailroom.
>>>>
>>>> To make it worse, buffer sizes are not actually uniform in idpf when
>>>> splitq is enabled, as there are several buffer queues, so rxq->rx_buf_size
>>>> is meaningless in this case.
>>>>
>>>> Use rxq->truesize as a frag_size for singleq and truesize of the first bufq
>>>> in AF_XDP ZC, as there is only one. Disable growinf tail for regular
>>>> splitq.
>>>>
>>>> Fixes: ac8a861f632e ("idpf: prepare structures to support XDP")
>>>> Reviewed-by: Aleksandr Loktionov <aleksandr.loktionov@xxxxxxxxx>
>>>> Signed-off-by: Larysa Zaremba <larysa.zaremba@xxxxxxxxx>
>>>> ---
>>>> drivers/net/ethernet/intel/idpf/xdp.c | 8 +++++++-
>>>> drivers/net/ethernet/intel/idpf/xsk.c | 1 +
>>>> 2 files changed, 8 insertions(+), 1 deletion(-)
>>>>
>>>> diff --git a/drivers/net/ethernet/intel/idpf/xdp.c b/drivers/net/ethernet/intel/idpf/xdp.c
>>>> index 958d16f87424..a152c9a26976 100644
>>>> --- a/drivers/net/ethernet/intel/idpf/xdp.c
>>>> +++ b/drivers/net/ethernet/intel/idpf/xdp.c
>>>> @@ -46,11 +46,17 @@ static int __idpf_xdp_rxq_info_init(struct idpf_rx_queue *rxq, void *arg)
>>>> {
>>>> const struct idpf_vport *vport = rxq->q_vector->vport;
>>>> bool split = idpf_is_queue_model_split(vport->rxq_model);
>>>> + u32 frag_size = 0;
>>>> int err;
>>>>
>>>> + if (idpf_queue_has(XSK, rxq) && split)
>>>> + frag_size = rxq->bufq_sets[0].bufq.truesize;
>>>> + else if (!split)
>>>> + frag_size = rxq->truesize;
>>>
>>> XDP and XSk are supported only in mode splitq mode, so you can remove
>>> the second condition and change the first one to just `has(XSK)`.
>>>
>
> But the function itself handles singleq case. I do not see why frag_size should
> be treated differently.

singleq will still be handled and will always receive 0 here as in case
when XSk is not enabled.

>
> Not that I am against of removing this logic, it would look more neat without
> these conditions.
>
>>>> +
>>>> err = __xdp_rxq_info_reg(&rxq->xdp_rxq, vport->netdev, rxq->idx,
>>>> rxq->q_vector->napi.napi_id,
>>>> - rxq->rx_buf_size);
>>>> + frag_size);
>>>> if (err)
>>>> return err;
>>>>
>>>> diff --git a/drivers/net/ethernet/intel/idpf/xsk.c b/drivers/net/ethernet/intel/idpf/xsk.c
>>>> index fd2cc43ab43c..febe1073b9b4 100644
>>>> --- a/drivers/net/ethernet/intel/idpf/xsk.c
>>>> +++ b/drivers/net/ethernet/intel/idpf/xsk.c
>>>> @@ -401,6 +401,7 @@ int idpf_xskfq_init(struct idpf_buf_queue *bufq)
>>>> bufq->pending = fq.pending;
>>>> bufq->thresh = fq.thresh;
>>>> bufq->rx_buf_size = fq.buf_len;
>>>> + bufq->truesize = xsk_pool_get_rx_frag_step(fq.pool);
>>
>> Better to do that in libeth_xdp rather than here?
>>
>
> Smth like that?
>
> diff --git a/drivers/net/ethernet/intel/idpf/xsk.c b/drivers/net/ethernet/intel/idpf/xsk.c
> index febe1073b9b4..de87455b92d7 100644
> --- a/drivers/net/ethernet/intel/idpf/xsk.c
> +++ b/drivers/net/ethernet/intel/idpf/xsk.c
> @@ -401,7 +401,7 @@ int idpf_xskfq_init(struct idpf_buf_queue *bufq)
> bufq->pending = fq.pending;
> bufq->thresh = fq.thresh;
> bufq->rx_buf_size = fq.buf_len;
> - bufq->truesize = xsk_pool_get_rx_frag_step(fq.pool);
> + bufq->truesize = fq.chunk_align;
>
> if (!idpf_xskfq_refill(bufq))
> netdev_err(bufq->pool->netdev,
> diff --git a/drivers/net/ethernet/intel/libeth/xsk.c b/drivers/net/ethernet/intel/libeth/xsk.c
> index 846e902e31b6..5b298558ecfd 100644
> --- a/drivers/net/ethernet/intel/libeth/xsk.c
> +++ b/drivers/net/ethernet/intel/libeth/xsk.c
> @@ -167,6 +167,7 @@ int libeth_xskfq_create(struct libeth_xskfq *fq)
> fq->pending = fq->count;
> fq->thresh = libeth_xdp_queue_threshold(fq->count);
> fq->buf_len = xsk_pool_get_rx_frame_size(fq->pool);
> + fq->chunk_align = xsk_pool_get_rx_frag_step(fq->pool);
>
> return 0;
> }
> diff --git a/include/net/libeth/xsk.h b/include/net/libeth/xsk.h
> index 481a7b28e6f2..a3ea90d30d17 100644
> --- a/include/net/libeth/xsk.h
> +++ b/include/net/libeth/xsk.h
> @@ -598,6 +598,7 @@ __libeth_xsk_run_pass(struct libeth_xdp_buff *xdp,
> * @thresh: threshold below which the queue is refilled
> * @buf_len: HW-writeable length per each buffer
> * @nid: ID of the closest NUMA node with memory
> + * @chunk_align: step between consecutive buffers, 0 if none exists
> */
> struct libeth_xskfq {
> struct_group_tagged(libeth_xskfq_fp, fp,
> @@ -615,6 +616,8 @@ struct libeth_xskfq {
>
> u32 buf_len;
> int nid;
> +
> + u32 chunk_align;
> };
>
> int libeth_xskfq_create(struct libeth_xskfq *fq);

Ah, I forgot that for XSkFQ we have a separate structure...

I'd do it like that:

u32 buf_len;
u32 truesize; // new field

int nid;
};

Maybe it's not really truesize, but it will be more straightforward and
clear to the users what it means and which field they should assign the
value to.
truesize/chunk_align is related closely to buf_len, so I'd group them
together in the struct, while nid is a different story, so it might stay
separated by a newline.

>
>
>>>>
>>>> if (!idpf_xskfq_refill(bufq))
>>>> netdev_err(bufq->pool->netdev,
Thanks,
Olek