Commit 96631ed1 authored by Eric Dumazet's avatar Eric Dumazet Committed by David S. Miller

udp: introduce sk_for_each_rcu_safenext()

Corey Minyard found a race added in commit 271b72c7
(udp: RCU handling for Unicast packets.)

 "If the socket is moved from one list to another list in-between the
 time the hash is calculated and the next field is accessed, and the
 socket has moved to the end of the new list, the traversal will not
 complete properly on the list it should have, since the socket will
 be on the end of the new list and there's not a way to tell it's on a
 new list and restart the list traversal.  I think that this can be
 solved by pre-fetching the "next" field (with proper barriers) before
 checking the hash."

This patch corrects this problem, introducing a new
sk_for_each_rcu_safenext() macro.
Signed-off-by: default avatarEric Dumazet <dada1@cosmosbay.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent f52b5054
...@@ -383,5 +383,22 @@ static inline void hlist_add_after_rcu(struct hlist_node *prev, ...@@ -383,5 +383,22 @@ static inline void hlist_add_after_rcu(struct hlist_node *prev,
({ tpos = hlist_entry(pos, typeof(*tpos), member); 1; }); \ ({ tpos = hlist_entry(pos, typeof(*tpos), member); 1; }); \
pos = rcu_dereference(pos->next)) pos = rcu_dereference(pos->next))
/**
* hlist_for_each_entry_rcu_safenext - iterate over rcu list of given type
* @tpos: the type * to use as a loop cursor.
* @pos: the &struct hlist_node to use as a loop cursor.
* @head: the head for your list.
* @member: the name of the hlist_node within the struct.
* @next: the &struct hlist_node to use as a next cursor
*
* Special version of hlist_for_each_entry_rcu that make sure
* each next pointer is fetched before each iteration.
*/
#define hlist_for_each_entry_rcu_safenext(tpos, pos, head, member, next) \
for (pos = rcu_dereference((head)->first); \
pos && ({ next = pos->next; smp_rmb(); prefetch(next); 1; }) && \
({ tpos = hlist_entry(pos, typeof(*tpos), member); 1; }); \
pos = rcu_dereference(next))
#endif /* __KERNEL__ */ #endif /* __KERNEL__ */
#endif #endif
...@@ -419,8 +419,8 @@ static __inline__ void sk_add_bind_node(struct sock *sk, ...@@ -419,8 +419,8 @@ static __inline__ void sk_add_bind_node(struct sock *sk,
#define sk_for_each(__sk, node, list) \ #define sk_for_each(__sk, node, list) \
hlist_for_each_entry(__sk, node, list, sk_node) hlist_for_each_entry(__sk, node, list, sk_node)
#define sk_for_each_rcu(__sk, node, list) \ #define sk_for_each_rcu_safenext(__sk, node, list, next) \
hlist_for_each_entry_rcu(__sk, node, list, sk_node) hlist_for_each_entry_rcu_safenext(__sk, node, list, sk_node, next)
#define sk_for_each_from(__sk, node) \ #define sk_for_each_from(__sk, node) \
if (__sk && ({ node = &(__sk)->sk_node; 1; })) \ if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
hlist_for_each_entry_from(__sk, node, sk_node) hlist_for_each_entry_from(__sk, node, sk_node)
......
...@@ -256,7 +256,7 @@ static struct sock *__udp4_lib_lookup(struct net *net, __be32 saddr, ...@@ -256,7 +256,7 @@ static struct sock *__udp4_lib_lookup(struct net *net, __be32 saddr,
int dif, struct udp_table *udptable) int dif, struct udp_table *udptable)
{ {
struct sock *sk, *result; struct sock *sk, *result;
struct hlist_node *node; struct hlist_node *node, *next;
unsigned short hnum = ntohs(dport); unsigned short hnum = ntohs(dport);
unsigned int hash = udp_hashfn(net, hnum); unsigned int hash = udp_hashfn(net, hnum);
struct udp_hslot *hslot = &udptable->hash[hash]; struct udp_hslot *hslot = &udptable->hash[hash];
...@@ -266,7 +266,7 @@ static struct sock *__udp4_lib_lookup(struct net *net, __be32 saddr, ...@@ -266,7 +266,7 @@ static struct sock *__udp4_lib_lookup(struct net *net, __be32 saddr,
begin: begin:
result = NULL; result = NULL;
badness = -1; badness = -1;
sk_for_each_rcu(sk, node, &hslot->head) { sk_for_each_rcu_safenext(sk, node, &hslot->head, next) {
/* /*
* lockless reader, and SLAB_DESTROY_BY_RCU items: * lockless reader, and SLAB_DESTROY_BY_RCU items:
* We must check this item was not moved to another chain * We must check this item was not moved to another chain
......
...@@ -98,7 +98,7 @@ static struct sock *__udp6_lib_lookup(struct net *net, ...@@ -98,7 +98,7 @@ static struct sock *__udp6_lib_lookup(struct net *net,
int dif, struct udp_table *udptable) int dif, struct udp_table *udptable)
{ {
struct sock *sk, *result; struct sock *sk, *result;
struct hlist_node *node; struct hlist_node *node, *next;
unsigned short hnum = ntohs(dport); unsigned short hnum = ntohs(dport);
unsigned int hash = udp_hashfn(net, hnum); unsigned int hash = udp_hashfn(net, hnum);
struct udp_hslot *hslot = &udptable->hash[hash]; struct udp_hslot *hslot = &udptable->hash[hash];
...@@ -108,7 +108,7 @@ static struct sock *__udp6_lib_lookup(struct net *net, ...@@ -108,7 +108,7 @@ static struct sock *__udp6_lib_lookup(struct net *net,
begin: begin:
result = NULL; result = NULL;
badness = -1; badness = -1;
sk_for_each_rcu(sk, node, &hslot->head) { sk_for_each_rcu_safenext(sk, node, &hslot->head, next) {
/* /*
* lockless reader, and SLAB_DESTROY_BY_RCU items: * lockless reader, and SLAB_DESTROY_BY_RCU items:
* We must check this item was not moved to another chain * We must check this item was not moved to another chain
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment