mirror of
https://github.com/torvalds/linux.git
synced 2025-12-07 20:06:24 +00:00
xsk: Use xsk_buff_pool directly for cq functions
Currently xsk_cq_{reserve_addr,submit,cancel}_locked() take xdp_sock as
an input argument but it is only used for pulling out xsk_buff_pool
pointer from it.
Change mentioned functions to take pool pointer as an input argument to
avoid unnecessary dereferences.
Signed-off-by: Maciej Fijalkowski <maciej.fijalkowski@intel.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Magnus Karlsson <magnus.karlsson@intel.com>
Link: https://lore.kernel.org/bpf/20241007122458.282590-7-maciej.fijalkowski@intel.com
This commit is contained in:
committed by
Daniel Borkmann
parent
1d10b2bed2
commit
e6c4047f51
@@ -527,34 +527,34 @@ static int xsk_wakeup(struct xdp_sock *xs, u8 flags)
|
|||||||
return dev->netdev_ops->ndo_xsk_wakeup(dev, xs->queue_id, flags);
|
return dev->netdev_ops->ndo_xsk_wakeup(dev, xs->queue_id, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int xsk_cq_reserve_addr_locked(struct xdp_sock *xs, u64 addr)
|
static int xsk_cq_reserve_addr_locked(struct xsk_buff_pool *pool, u64 addr)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
spin_lock_irqsave(&xs->pool->cq_lock, flags);
|
spin_lock_irqsave(&pool->cq_lock, flags);
|
||||||
ret = xskq_prod_reserve_addr(xs->pool->cq, addr);
|
ret = xskq_prod_reserve_addr(pool->cq, addr);
|
||||||
spin_unlock_irqrestore(&xs->pool->cq_lock, flags);
|
spin_unlock_irqrestore(&pool->cq_lock, flags);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void xsk_cq_submit_locked(struct xdp_sock *xs, u32 n)
|
static void xsk_cq_submit_locked(struct xsk_buff_pool *pool, u32 n)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
spin_lock_irqsave(&xs->pool->cq_lock, flags);
|
spin_lock_irqsave(&pool->cq_lock, flags);
|
||||||
xskq_prod_submit_n(xs->pool->cq, n);
|
xskq_prod_submit_n(pool->cq, n);
|
||||||
spin_unlock_irqrestore(&xs->pool->cq_lock, flags);
|
spin_unlock_irqrestore(&pool->cq_lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void xsk_cq_cancel_locked(struct xdp_sock *xs, u32 n)
|
static void xsk_cq_cancel_locked(struct xsk_buff_pool *pool, u32 n)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
spin_lock_irqsave(&xs->pool->cq_lock, flags);
|
spin_lock_irqsave(&pool->cq_lock, flags);
|
||||||
xskq_prod_cancel_n(xs->pool->cq, n);
|
xskq_prod_cancel_n(pool->cq, n);
|
||||||
spin_unlock_irqrestore(&xs->pool->cq_lock, flags);
|
spin_unlock_irqrestore(&pool->cq_lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static u32 xsk_get_num_desc(struct sk_buff *skb)
|
static u32 xsk_get_num_desc(struct sk_buff *skb)
|
||||||
@@ -571,7 +571,7 @@ static void xsk_destruct_skb(struct sk_buff *skb)
|
|||||||
*compl->tx_timestamp = ktime_get_tai_fast_ns();
|
*compl->tx_timestamp = ktime_get_tai_fast_ns();
|
||||||
}
|
}
|
||||||
|
|
||||||
xsk_cq_submit_locked(xdp_sk(skb->sk), xsk_get_num_desc(skb));
|
xsk_cq_submit_locked(xdp_sk(skb->sk)->pool, xsk_get_num_desc(skb));
|
||||||
sock_wfree(skb);
|
sock_wfree(skb);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -587,7 +587,7 @@ static void xsk_consume_skb(struct sk_buff *skb)
|
|||||||
struct xdp_sock *xs = xdp_sk(skb->sk);
|
struct xdp_sock *xs = xdp_sk(skb->sk);
|
||||||
|
|
||||||
skb->destructor = sock_wfree;
|
skb->destructor = sock_wfree;
|
||||||
xsk_cq_cancel_locked(xs, xsk_get_num_desc(skb));
|
xsk_cq_cancel_locked(xs->pool, xsk_get_num_desc(skb));
|
||||||
/* Free skb without triggering the perf drop trace */
|
/* Free skb without triggering the perf drop trace */
|
||||||
consume_skb(skb);
|
consume_skb(skb);
|
||||||
xs->skb = NULL;
|
xs->skb = NULL;
|
||||||
@@ -765,7 +765,7 @@ free_err:
|
|||||||
xskq_cons_release(xs->tx);
|
xskq_cons_release(xs->tx);
|
||||||
} else {
|
} else {
|
||||||
/* Let application retry */
|
/* Let application retry */
|
||||||
xsk_cq_cancel_locked(xs, 1);
|
xsk_cq_cancel_locked(xs->pool, 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
return ERR_PTR(err);
|
return ERR_PTR(err);
|
||||||
@@ -802,7 +802,7 @@ static int __xsk_generic_xmit(struct sock *sk)
|
|||||||
* if there is space in it. This avoids having to implement
|
* if there is space in it. This avoids having to implement
|
||||||
* any buffering in the Tx path.
|
* any buffering in the Tx path.
|
||||||
*/
|
*/
|
||||||
if (xsk_cq_reserve_addr_locked(xs, desc.addr))
|
if (xsk_cq_reserve_addr_locked(xs->pool, desc.addr))
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
skb = xsk_build_skb(xs, &desc);
|
skb = xsk_build_skb(xs, &desc);
|
||||||
|
|||||||
Reference in New Issue
Block a user