Commit 8b80cda5 authored by Eliezer Tamir's avatar Eliezer Tamir Committed by David S. Miller
Browse files

net: rename ll methods to busy-poll



Rename ndo_ll_poll to ndo_busy_poll.
Rename sk_mark_ll to sk_mark_napi_id.
Rename skb_mark_ll to skb_mark_napi_id.
Correct all useres of these functions.
Update comments and defines  in include/net/busy_poll.h
Signed-off-by: default avatarEliezer Tamir <eliezer.tamir@linux.intel.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 076bb0c8
...@@ -990,7 +990,7 @@ reuse_rx: ...@@ -990,7 +990,7 @@ reuse_rx:
__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q),
le16_to_cpu(cqe_fp->vlan_tag)); le16_to_cpu(cqe_fp->vlan_tag));
skb_mark_ll(skb, &fp->napi); skb_mark_napi_id(skb, &fp->napi);
if (bnx2x_fp_ll_polling(fp)) if (bnx2x_fp_ll_polling(fp))
netif_receive_skb(skb); netif_receive_skb(skb);
......
...@@ -12027,7 +12027,7 @@ static const struct net_device_ops bnx2x_netdev_ops = { ...@@ -12027,7 +12027,7 @@ static const struct net_device_ops bnx2x_netdev_ops = {
#endif #endif
#ifdef CONFIG_NET_LL_RX_POLL #ifdef CONFIG_NET_LL_RX_POLL
.ndo_ll_poll = bnx2x_low_latency_recv, .ndo_busy_poll = bnx2x_low_latency_recv,
#endif #endif
}; };
......
...@@ -1978,7 +1978,7 @@ static int ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector, ...@@ -1978,7 +1978,7 @@ static int ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector,
} }
#endif /* IXGBE_FCOE */ #endif /* IXGBE_FCOE */
skb_mark_ll(skb, &q_vector->napi); skb_mark_napi_id(skb, &q_vector->napi);
ixgbe_rx_skb(q_vector, skb); ixgbe_rx_skb(q_vector, skb);
/* update budget accounting */ /* update budget accounting */
...@@ -7228,7 +7228,7 @@ static const struct net_device_ops ixgbe_netdev_ops = { ...@@ -7228,7 +7228,7 @@ static const struct net_device_ops ixgbe_netdev_ops = {
.ndo_poll_controller = ixgbe_netpoll, .ndo_poll_controller = ixgbe_netpoll,
#endif #endif
#ifdef CONFIG_NET_LL_RX_POLL #ifdef CONFIG_NET_LL_RX_POLL
.ndo_ll_poll = ixgbe_low_latency_recv, .ndo_busy_poll = ixgbe_low_latency_recv,
#endif #endif
#ifdef IXGBE_FCOE #ifdef IXGBE_FCOE
.ndo_fcoe_ddp_setup = ixgbe_fcoe_ddp_get, .ndo_fcoe_ddp_setup = ixgbe_fcoe_ddp_get,
......
...@@ -2141,7 +2141,7 @@ static const struct net_device_ops mlx4_netdev_ops = { ...@@ -2141,7 +2141,7 @@ static const struct net_device_ops mlx4_netdev_ops = {
.ndo_rx_flow_steer = mlx4_en_filter_rfs, .ndo_rx_flow_steer = mlx4_en_filter_rfs,
#endif #endif
#ifdef CONFIG_NET_LL_RX_POLL #ifdef CONFIG_NET_LL_RX_POLL
.ndo_ll_poll = mlx4_en_low_latency_recv, .ndo_busy_poll = mlx4_en_low_latency_recv,
#endif #endif
}; };
......
...@@ -767,7 +767,7 @@ int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int bud ...@@ -767,7 +767,7 @@ int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int bud
timestamp); timestamp);
} }
skb_mark_ll(skb, &cq->napi); skb_mark_napi_id(skb, &cq->napi);
/* Push it up the stack */ /* Push it up the stack */
netif_receive_skb(skb); netif_receive_skb(skb);
......
...@@ -974,7 +974,7 @@ struct net_device_ops { ...@@ -974,7 +974,7 @@ struct net_device_ops {
void (*ndo_netpoll_cleanup)(struct net_device *dev); void (*ndo_netpoll_cleanup)(struct net_device *dev);
#endif #endif
#ifdef CONFIG_NET_LL_RX_POLL #ifdef CONFIG_NET_LL_RX_POLL
int (*ndo_ll_poll)(struct napi_struct *dev); int (*ndo_busy_poll)(struct napi_struct *dev);
#endif #endif
int (*ndo_set_vf_mac)(struct net_device *dev, int (*ndo_set_vf_mac)(struct net_device *dev,
int queue, u8 *mac); int queue, u8 *mac);
......
/* /*
* Low Latency Sockets * net busy poll support
* Copyright(c) 2013 Intel Corporation. * Copyright(c) 2013 Intel Corporation.
* *
* This program is free software; you can redistribute it and/or modify it * This program is free software; you can redistribute it and/or modify it
...@@ -21,8 +21,8 @@ ...@@ -21,8 +21,8 @@
* e1000-devel Mailing List <e1000-devel@lists.sourceforge.net> * e1000-devel Mailing List <e1000-devel@lists.sourceforge.net>
*/ */
#ifndef _LINUX_NET_LL_POLL_H #ifndef _LINUX_NET_BUSY_POLL_H
#define _LINUX_NET_LL_POLL_H #define _LINUX_NET_BUSY_POLL_H
#include <linux/netdevice.h> #include <linux/netdevice.h>
#include <net/ip.h> #include <net/ip.h>
...@@ -110,11 +110,11 @@ static inline bool sk_busy_loop(struct sock *sk, int nonblock) ...@@ -110,11 +110,11 @@ static inline bool sk_busy_loop(struct sock *sk, int nonblock)
goto out; goto out;
ops = napi->dev->netdev_ops; ops = napi->dev->netdev_ops;
if (!ops->ndo_ll_poll) if (!ops->ndo_busy_poll)
goto out; goto out;
do { do {
rc = ops->ndo_ll_poll(napi); rc = ops->ndo_busy_poll(napi);
if (rc == LL_FLUSH_FAILED) if (rc == LL_FLUSH_FAILED)
break; /* permanent failure */ break; /* permanent failure */
...@@ -134,13 +134,14 @@ out: ...@@ -134,13 +134,14 @@ out:
} }
/* used in the NIC receive handler to mark the skb */ /* used in the NIC receive handler to mark the skb */
static inline void skb_mark_ll(struct sk_buff *skb, struct napi_struct *napi) static inline void skb_mark_napi_id(struct sk_buff *skb,
struct napi_struct *napi)
{ {
skb->napi_id = napi->napi_id; skb->napi_id = napi->napi_id;
} }
/* used in the protocol hanlder to propagate the napi_id to the socket */ /* used in the protocol hanlder to propagate the napi_id to the socket */
static inline void sk_mark_ll(struct sock *sk, struct sk_buff *skb) static inline void sk_mark_napi_id(struct sock *sk, struct sk_buff *skb)
{ {
sk->sk_napi_id = skb->napi_id; sk->sk_napi_id = skb->napi_id;
} }
...@@ -166,11 +167,12 @@ static inline bool sk_busy_poll(struct sock *sk, int nonblock) ...@@ -166,11 +167,12 @@ static inline bool sk_busy_poll(struct sock *sk, int nonblock)
return false; return false;
} }
static inline void skb_mark_ll(struct sk_buff *skb, struct napi_struct *napi) static inline void skb_mark_napi_id(struct sk_buff *skb,
struct napi_struct *napi)
{ {
} }
static inline void sk_mark_ll(struct sock *sk, struct sk_buff *skb) static inline void sk_mark_napi_id(struct sock *sk, struct sk_buff *skb)
{ {
} }
...@@ -180,4 +182,4 @@ static inline bool busy_loop_timeout(unsigned long end_time) ...@@ -180,4 +182,4 @@ static inline bool busy_loop_timeout(unsigned long end_time)
} }
#endif /* CONFIG_NET_LL_RX_POLL */ #endif /* CONFIG_NET_LL_RX_POLL */
#endif /* _LINUX_NET_LL_POLL_H */ #endif /* _LINUX_NET_BUSY_POLL_H */
...@@ -1994,7 +1994,7 @@ process: ...@@ -1994,7 +1994,7 @@ process:
if (sk_filter(sk, skb)) if (sk_filter(sk, skb))
goto discard_and_relse; goto discard_and_relse;
sk_mark_ll(sk, skb); sk_mark_napi_id(sk, skb);
skb->dev = NULL; skb->dev = NULL;
bh_lock_sock_nested(sk); bh_lock_sock_nested(sk);
......
...@@ -1713,7 +1713,7 @@ int __udp4_lib_rcv(struct sk_buff *skb, struct udp_table *udptable, ...@@ -1713,7 +1713,7 @@ int __udp4_lib_rcv(struct sk_buff *skb, struct udp_table *udptable,
if (sk != NULL) { if (sk != NULL) {
int ret; int ret;
sk_mark_ll(sk, skb); sk_mark_napi_id(sk, skb);
ret = udp_queue_rcv_skb(sk, skb); ret = udp_queue_rcv_skb(sk, skb);
sock_put(sk); sock_put(sk);
......
...@@ -1499,7 +1499,7 @@ process: ...@@ -1499,7 +1499,7 @@ process:
if (sk_filter(sk, skb)) if (sk_filter(sk, skb))
goto discard_and_relse; goto discard_and_relse;
sk_mark_ll(sk, skb); sk_mark_napi_id(sk, skb);
skb->dev = NULL; skb->dev = NULL;
bh_lock_sock_nested(sk); bh_lock_sock_nested(sk);
......
...@@ -844,7 +844,7 @@ int __udp6_lib_rcv(struct sk_buff *skb, struct udp_table *udptable, ...@@ -844,7 +844,7 @@ int __udp6_lib_rcv(struct sk_buff *skb, struct udp_table *udptable,
if (sk != NULL) { if (sk != NULL) {
int ret; int ret;
sk_mark_ll(sk, skb); sk_mark_napi_id(sk, skb);
ret = udpv6_queue_rcv_skb(sk, skb); ret = udpv6_queue_rcv_skb(sk, skb);
sock_put(sk); sock_put(sk);
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment