it's a Per-CPU variable.
Default: 64
-low_latency_read
+busy_read
----------------
Low latency busy poll timeout for socket reads. (needs CONFIG_NET_LL_RX_POLL)
Approximate time in us to busy loop waiting for packets on the device queue.
-This sets the default value of the SO_LL socket option.
-Can be set or overridden per socket by setting socket option SO_LL, which is
-the preferred method of enabling.
-If you need to enable the feature globally via sysctl, a value of 50 is recommended.
+This sets the default value of the SO_BUSY_POLL socket option.
+Can be set or overridden per socket by setting socket option SO_BUSY_POLL,
+which is the preferred method of enabling. If you need to enable the feature
+globally via sysctl, a value of 50 is recommended.
Will increase power usage.
Default: 0 (off)
-low_latency_poll
+busy_poll
----------------
Low latency busy poll timeout for poll and select. (needs CONFIG_NET_LL_RX_POLL)
Approximate time in us to busy loop waiting for events.
Recommended value depends on the number of sockets you poll on.
For several sockets 50, for several hundreds 100.
For more than that you probably want to use epoll.
-Note that only sockets with SO_LL set will be busy polled, so you want to either
-selectively set SO_LL on those sockets or set sysctl.net.low_latency_read globally.
+Note that only sockets with SO_BUSY_POLL set will be busy polled,
+so you want to either selectively set SO_BUSY_POLL on those sockets or set
+sysctl.net.busy_read globally.
Will increase power usage.
Default: 0 (off)
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _UAPI_ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* __ASM_AVR32_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_IA64_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_M32R_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _UAPI_ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 0x4026
-#define SO_LL 0x4027
+#define SO_BUSY_POLL 0x4027
/* O_NONBLOCK clashes with the bits used for socket types. Therefore we
* have to define SOCK_NONBLOCK to a different value here.
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_POWERPC_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _ASM_SOCKET_H */
#define SO_SELECT_ERR_QUEUE 0x0029
-#define SO_LL 0x0030
+#define SO_BUSY_POLL 0x0030
/* Security levels - as per NRL IPv6 - don't actually do anything */
#define SO_SECURITY_AUTHENTICATION 0x5001
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* _XTENSA_SOCKET_H */
#include <net/tcp.h>
#include <net/ipv6.h>
#include <net/ip6_checksum.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <linux/prefetch.h>
#include "bnx2x_cmn.h"
#include "bnx2x_init.h"
__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q),
le16_to_cpu(cqe_fp->vlan_tag));
- skb_mark_ll(skb, &fp->napi);
+ skb_mark_napi_id(skb, &fp->napi);
if (bnx2x_fp_ll_polling(fp))
netif_receive_skb(skb);
#endif
#ifdef CONFIG_NET_LL_RX_POLL
- .ndo_ll_poll = bnx2x_low_latency_recv,
+ .ndo_busy_poll = bnx2x_low_latency_recv,
#endif
};
#include <linux/dca.h>
#endif
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#ifdef CONFIG_NET_LL_RX_POLL
#define LL_EXTENDED_STATS
}
#endif /* IXGBE_FCOE */
- skb_mark_ll(skb, &q_vector->napi);
+ skb_mark_napi_id(skb, &q_vector->napi);
ixgbe_rx_skb(q_vector, skb);
/* update budget accounting */
.ndo_poll_controller = ixgbe_netpoll,
#endif
#ifdef CONFIG_NET_LL_RX_POLL
- .ndo_ll_poll = ixgbe_low_latency_recv,
+ .ndo_busy_poll = ixgbe_low_latency_recv,
#endif
#ifdef IXGBE_FCOE
.ndo_fcoe_ddp_setup = ixgbe_fcoe_ddp_get,
#include <linux/slab.h>
#include <linux/hash.h>
#include <net/ip.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <linux/mlx4/driver.h>
#include <linux/mlx4/device.h>
.ndo_rx_flow_steer = mlx4_en_filter_rfs,
#endif
#ifdef CONFIG_NET_LL_RX_POLL
- .ndo_ll_poll = mlx4_en_low_latency_recv,
+ .ndo_busy_poll = mlx4_en_low_latency_recv,
#endif
};
*
*/
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <linux/mlx4/cq.h>
#include <linux/slab.h>
#include <linux/mlx4/qp.h>
timestamp);
}
- skb_mark_ll(skb, &cq->napi);
+ skb_mark_napi_id(skb, &cq->napi);
/* Push it up the stack */
netif_receive_skb(skb);
#include <linux/hrtimer.h>
#include <linux/sched/rt.h>
#include <linux/freezer.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <asm/uaccess.h>
void (*ndo_netpoll_cleanup)(struct net_device *dev);
#endif
#ifdef CONFIG_NET_LL_RX_POLL
- int (*ndo_ll_poll)(struct napi_struct *dev);
+ int (*ndo_busy_poll)(struct napi_struct *dev);
#endif
int (*ndo_set_vf_mac)(struct net_device *dev,
int queue, u8 *mac);
/*
- * Low Latency Sockets
+ * net busy poll support
* Copyright(c) 2013 Intel Corporation.
*
* This program is free software; you can redistribute it and/or modify it
* e1000-devel Mailing List <e1000-devel@lists.sourceforge.net>
*/
-#ifndef _LINUX_NET_LL_POLL_H
-#define _LINUX_NET_LL_POLL_H
+#ifndef _LINUX_NET_BUSY_POLL_H
+#define _LINUX_NET_BUSY_POLL_H
#include <linux/netdevice.h>
#include <net/ip.h>
#ifdef CONFIG_NET_LL_RX_POLL
struct napi_struct;
-extern unsigned int sysctl_net_ll_read __read_mostly;
-extern unsigned int sysctl_net_ll_poll __read_mostly;
+extern unsigned int sysctl_net_busy_read __read_mostly;
+extern unsigned int sysctl_net_busy_poll __read_mostly;
/* return values from ndo_ll_poll */
#define LL_FLUSH_FAILED -1
static inline bool net_busy_loop_on(void)
{
- return sysctl_net_ll_poll;
+ return sysctl_net_busy_poll;
}
/* a wrapper to make debug_smp_processor_id() happy
/* in poll/select we use the global sysctl_net_ll_poll value */
static inline unsigned long busy_loop_end_time(void)
{
- return busy_loop_us_clock() + ACCESS_ONCE(sysctl_net_ll_poll);
+ return busy_loop_us_clock() + ACCESS_ONCE(sysctl_net_busy_poll);
}
static inline bool sk_can_busy_loop(struct sock *sk)
goto out;
ops = napi->dev->netdev_ops;
- if (!ops->ndo_ll_poll)
+ if (!ops->ndo_busy_poll)
goto out;
do {
- rc = ops->ndo_ll_poll(napi);
+ rc = ops->ndo_busy_poll(napi);
if (rc == LL_FLUSH_FAILED)
break; /* permanent failure */
}
/* used in the NIC receive handler to mark the skb */
-static inline void skb_mark_ll(struct sk_buff *skb, struct napi_struct *napi)
+static inline void skb_mark_napi_id(struct sk_buff *skb,
+ struct napi_struct *napi)
{
skb->napi_id = napi->napi_id;
}
/* used in the protocol hanlder to propagate the napi_id to the socket */
-static inline void sk_mark_ll(struct sock *sk, struct sk_buff *skb)
+static inline void sk_mark_napi_id(struct sock *sk, struct sk_buff *skb)
{
sk->sk_napi_id = skb->napi_id;
}
return false;
}
-static inline void skb_mark_ll(struct sk_buff *skb, struct napi_struct *napi)
+static inline void skb_mark_napi_id(struct sk_buff *skb,
+ struct napi_struct *napi)
{
}
-static inline void sk_mark_ll(struct sock *sk, struct sk_buff *skb)
+static inline void sk_mark_napi_id(struct sock *sk, struct sk_buff *skb)
{
}
}
#endif /* CONFIG_NET_LL_RX_POLL */
-#endif /* _LINUX_NET_LL_POLL_H */
+#endif /* _LINUX_NET_BUSY_POLL_H */
#define SO_SELECT_ERR_QUEUE 45
-#define SO_LL 46
+#define SO_BUSY_POLL 46
#endif /* __ASM_GENERIC_SOCKET_H */
#include <net/sock.h>
#include <net/tcp_states.h>
#include <trace/events/skb.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
/*
* Is a socket 'connection oriented' ?
#include <net/tcp.h>
#endif
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
static DEFINE_MUTEX(proto_list_mutex);
static LIST_HEAD(proto_list);
break;
#ifdef CONFIG_NET_LL_RX_POLL
- case SO_LL:
+ case SO_BUSY_POLL:
/* allow unprivileged users to decrease the value */
if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN))
ret = -EPERM;
break;
#ifdef CONFIG_NET_LL_RX_POLL
- case SO_LL:
+ case SO_BUSY_POLL:
v.val = sk->sk_ll_usec;
break;
#endif
#ifdef CONFIG_NET_LL_RX_POLL
sk->sk_napi_id = 0;
- sk->sk_ll_usec = sysctl_net_ll_read;
+ sk->sk_ll_usec = sysctl_net_busy_read;
#endif
/*
#include <net/ip.h>
#include <net/sock.h>
#include <net/net_ratelimit.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
static int one = 1;
#endif /* CONFIG_NET_FLOW_LIMIT */
#ifdef CONFIG_NET_LL_RX_POLL
{
- .procname = "low_latency_poll",
- .data = &sysctl_net_ll_poll,
+ .procname = "busy_poll",
+ .data = &sysctl_net_busy_poll,
.maxlen = sizeof(unsigned int),
.mode = 0644,
.proc_handler = proc_dointvec
},
{
- .procname = "low_latency_read",
- .data = &sysctl_net_ll_read,
+ .procname = "busy_read",
+ .data = &sysctl_net_busy_read,
.maxlen = sizeof(unsigned int),
.mode = 0644,
.proc_handler = proc_dointvec
#include <asm/uaccess.h>
#include <asm/ioctls.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
int sysctl_tcp_fin_timeout __read_mostly = TCP_FIN_TIMEOUT;
#include <net/netdma.h>
#include <net/secure_seq.h>
#include <net/tcp_memcontrol.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <linux/inet.h>
#include <linux/ipv6.h>
if (sk_filter(sk, skb))
goto discard_and_relse;
- sk_mark_ll(sk, skb);
+ sk_mark_napi_id(sk, skb);
skb->dev = NULL;
bh_lock_sock_nested(sk);
#include <trace/events/udp.h>
#include <linux/static_key.h>
#include <trace/events/skb.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include "udp_impl.h"
struct udp_table udp_table __read_mostly;
if (sk != NULL) {
int ret;
- sk_mark_ll(sk, skb);
+ sk_mark_napi_id(sk, skb);
ret = udp_queue_rcv_skb(sk, skb);
sock_put(sk);
#include <net/inet_common.h>
#include <net/secure_seq.h>
#include <net/tcp_memcontrol.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <asm/uaccess.h>
if (sk_filter(sk, skb))
goto discard_and_relse;
- sk_mark_ll(sk, skb);
+ sk_mark_napi_id(sk, skb);
skb->dev = NULL;
bh_lock_sock_nested(sk);
#include <net/ip6_checksum.h>
#include <net/xfrm.h>
#include <net/inet6_hashtables.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#include <linux/proc_fs.h>
#include <linux/seq_file.h>
if (sk != NULL) {
int ret;
- sk_mark_ll(sk, skb);
+ sk_mark_napi_id(sk, skb);
ret = udpv6_queue_rcv_skb(sk, skb);
sock_put(sk);
#include <linux/route.h>
#include <linux/sockios.h>
#include <linux/atalk.h>
-#include <net/ll_poll.h>
+#include <net/busy_poll.h>
#ifdef CONFIG_NET_LL_RX_POLL
-unsigned int sysctl_net_ll_read __read_mostly;
-unsigned int sysctl_net_ll_poll __read_mostly;
+unsigned int sysctl_net_busy_read __read_mostly;
+unsigned int sysctl_net_busy_poll __read_mostly;
#endif
static int sock_no_open(struct inode *irrelevant, struct file *dontcare);