Commit 2c1409a0 authored by Eric Dumazet's avatar Eric Dumazet Committed by David S. Miller

inetpeer: Optimize inet_getid()

While investigating for network latencies, I found inet_getid() was a
contention point for some workloads, as inet_peer_idlock is shared
by all inet_getid() users regardless of peers.

One way to fix this is to make ip_id_count an atomic_t instead
of __u16, and use atomic_add_return().

In order to keep sizeof(struct inet_peer) = 64 on 64bit arches
tcp_ts_stamp is also converted to __u32 instead of "unsigned long".
Signed-off-by: default avatarEric Dumazet <eric.dumazet@gmail.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 234b27c3
...@@ -17,15 +17,15 @@ struct inet_peer { ...@@ -17,15 +17,15 @@ struct inet_peer {
/* group together avl_left,avl_right,v4daddr to speedup lookups */ /* group together avl_left,avl_right,v4daddr to speedup lookups */
struct inet_peer *avl_left, *avl_right; struct inet_peer *avl_left, *avl_right;
__be32 v4daddr; /* peer's address */ __be32 v4daddr; /* peer's address */
__u16 avl_height; __u32 avl_height;
__u16 ip_id_count; /* IP ID for the next packet */
struct list_head unused; struct list_head unused;
__u32 dtime; /* the time of last use of not __u32 dtime; /* the time of last use of not
* referenced entries */ * referenced entries */
atomic_t refcnt; atomic_t refcnt;
atomic_t rid; /* Frag reception counter */ atomic_t rid; /* Frag reception counter */
atomic_t ip_id_count; /* IP ID for the next packet */
__u32 tcp_ts; __u32 tcp_ts;
unsigned long tcp_ts_stamp; __u32 tcp_ts_stamp;
}; };
void inet_initpeers(void) __init; void inet_initpeers(void) __init;
...@@ -36,17 +36,11 @@ struct inet_peer *inet_getpeer(__be32 daddr, int create); ...@@ -36,17 +36,11 @@ struct inet_peer *inet_getpeer(__be32 daddr, int create);
/* can be called from BH context or outside */ /* can be called from BH context or outside */
extern void inet_putpeer(struct inet_peer *p); extern void inet_putpeer(struct inet_peer *p);
extern spinlock_t inet_peer_idlock;
/* can be called with or without local BH being disabled */ /* can be called with or without local BH being disabled */
static inline __u16 inet_getid(struct inet_peer *p, int more) static inline __u16 inet_getid(struct inet_peer *p, int more)
{ {
__u16 id; more++;
return atomic_add_return(more, &p->ip_id_count) - more;
spin_lock_bh(&inet_peer_idlock);
id = p->ip_id_count;
p->ip_id_count += 1 + more;
spin_unlock_bh(&inet_peer_idlock);
return id;
} }
#endif /* _NET_INETPEER_H */ #endif /* _NET_INETPEER_H */
...@@ -67,9 +67,6 @@ ...@@ -67,9 +67,6 @@
* ip_id_count: idlock * ip_id_count: idlock
*/ */
/* Exported for inet_getid inline function. */
DEFINE_SPINLOCK(inet_peer_idlock);
static struct kmem_cache *peer_cachep __read_mostly; static struct kmem_cache *peer_cachep __read_mostly;
#define node_height(x) x->avl_height #define node_height(x) x->avl_height
...@@ -390,7 +387,7 @@ struct inet_peer *inet_getpeer(__be32 daddr, int create) ...@@ -390,7 +387,7 @@ struct inet_peer *inet_getpeer(__be32 daddr, int create)
n->v4daddr = daddr; n->v4daddr = daddr;
atomic_set(&n->refcnt, 1); atomic_set(&n->refcnt, 1);
atomic_set(&n->rid, 0); atomic_set(&n->rid, 0);
n->ip_id_count = secure_ip_id(daddr); atomic_set(&n->ip_id_count, secure_ip_id(daddr));
n->tcp_ts_stamp = 0; n->tcp_ts_stamp = 0;
write_lock_bh(&peer_pool_lock); write_lock_bh(&peer_pool_lock);
......
...@@ -2852,7 +2852,7 @@ static int rt_fill_info(struct net *net, ...@@ -2852,7 +2852,7 @@ static int rt_fill_info(struct net *net,
error = rt->u.dst.error; error = rt->u.dst.error;
expires = rt->u.dst.expires ? rt->u.dst.expires - jiffies : 0; expires = rt->u.dst.expires ? rt->u.dst.expires - jiffies : 0;
if (rt->peer) { if (rt->peer) {
id = rt->peer->ip_id_count; id = atomic_read(&rt->peer->ip_id_count) & 0xffff;
if (rt->peer->tcp_ts_stamp) { if (rt->peer->tcp_ts_stamp) {
ts = rt->peer->tcp_ts; ts = rt->peer->tcp_ts;
tsage = get_seconds() - rt->peer->tcp_ts_stamp; tsage = get_seconds() - rt->peer->tcp_ts_stamp;
......
...@@ -204,7 +204,7 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) ...@@ -204,7 +204,7 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len)
* when trying new connection. * when trying new connection.
*/ */
if (peer != NULL && if (peer != NULL &&
peer->tcp_ts_stamp + TCP_PAWS_MSL >= get_seconds()) { (u32)get_seconds() - peer->tcp_ts_stamp <= TCP_PAWS_MSL) {
tp->rx_opt.ts_recent_stamp = peer->tcp_ts_stamp; tp->rx_opt.ts_recent_stamp = peer->tcp_ts_stamp;
tp->rx_opt.ts_recent = peer->tcp_ts; tp->rx_opt.ts_recent = peer->tcp_ts;
} }
...@@ -1308,7 +1308,7 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) ...@@ -1308,7 +1308,7 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb)
tcp_death_row.sysctl_tw_recycle && tcp_death_row.sysctl_tw_recycle &&
(peer = rt_get_peer((struct rtable *)dst)) != NULL && (peer = rt_get_peer((struct rtable *)dst)) != NULL &&
peer->v4daddr == saddr) { peer->v4daddr == saddr) {
if (get_seconds() < peer->tcp_ts_stamp + TCP_PAWS_MSL && if ((u32)get_seconds() - peer->tcp_ts_stamp < TCP_PAWS_MSL &&
(s32)(peer->tcp_ts - req->ts_recent) > (s32)(peer->tcp_ts - req->ts_recent) >
TCP_PAWS_WINDOW) { TCP_PAWS_WINDOW) {
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_PAWSPASSIVEREJECTED); NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_PAWSPASSIVEREJECTED);
...@@ -1727,9 +1727,9 @@ int tcp_v4_remember_stamp(struct sock *sk) ...@@ -1727,9 +1727,9 @@ int tcp_v4_remember_stamp(struct sock *sk)
if (peer) { if (peer) {
if ((s32)(peer->tcp_ts - tp->rx_opt.ts_recent) <= 0 || if ((s32)(peer->tcp_ts - tp->rx_opt.ts_recent) <= 0 ||
(peer->tcp_ts_stamp + TCP_PAWS_MSL < get_seconds() && ((u32)get_seconds() - peer->tcp_ts_stamp > TCP_PAWS_MSL &&
peer->tcp_ts_stamp <= tp->rx_opt.ts_recent_stamp)) { peer->tcp_ts_stamp <= (u32)tp->rx_opt.ts_recent_stamp)) {
peer->tcp_ts_stamp = tp->rx_opt.ts_recent_stamp; peer->tcp_ts_stamp = (u32)tp->rx_opt.ts_recent_stamp;
peer->tcp_ts = tp->rx_opt.ts_recent; peer->tcp_ts = tp->rx_opt.ts_recent;
} }
if (release_it) if (release_it)
...@@ -1748,9 +1748,9 @@ int tcp_v4_tw_remember_stamp(struct inet_timewait_sock *tw) ...@@ -1748,9 +1748,9 @@ int tcp_v4_tw_remember_stamp(struct inet_timewait_sock *tw)
const struct tcp_timewait_sock *tcptw = tcp_twsk((struct sock *)tw); const struct tcp_timewait_sock *tcptw = tcp_twsk((struct sock *)tw);
if ((s32)(peer->tcp_ts - tcptw->tw_ts_recent) <= 0 || if ((s32)(peer->tcp_ts - tcptw->tw_ts_recent) <= 0 ||
(peer->tcp_ts_stamp + TCP_PAWS_MSL < get_seconds() && ((u32)get_seconds() - peer->tcp_ts_stamp > TCP_PAWS_MSL &&
peer->tcp_ts_stamp <= tcptw->tw_ts_recent_stamp)) { peer->tcp_ts_stamp <= (u32)tcptw->tw_ts_recent_stamp)) {
peer->tcp_ts_stamp = tcptw->tw_ts_recent_stamp; peer->tcp_ts_stamp = (u32)tcptw->tw_ts_recent_stamp;
peer->tcp_ts = tcptw->tw_ts_recent; peer->tcp_ts = tcptw->tw_ts_recent;
} }
inet_putpeer(peer); inet_putpeer(peer);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment