Commit 5dae7a69 authored by Dipankar Sarma's avatar Dipankar Sarma Committed by Linus Torvalds

[PATCH] rcu: abstracted RCU dereferencing

Use abstracted RCU API to dereference RCU protected data.  Hides barrier
details.  Patch from Paul McKenney.

This patch introduced an rcu_dereference() macro that replaces most uses of
smp_read_barrier_depends().  The new macro has the advantage of explicitly
documenting which pointers are protected by RCU -- in contrast, it is
sometimes difficult to figure out which pointer is being protected by a given
smp_read_barrier_depends() call.
Signed-off-by: default avatarPaul McKenney <paulmck@us.ibm.com>
Signed-off-by: default avatarAndrew Morton <akpm@osdl.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent 9711268c
...@@ -48,8 +48,7 @@ static void mce_log(struct mce *mce) ...@@ -48,8 +48,7 @@ static void mce_log(struct mce *mce)
mce->finished = 0; mce->finished = 0;
smp_wmb(); smp_wmb();
for (;;) { for (;;) {
entry = mcelog.next; entry = rcu_dereference(mcelog.next);
read_barrier_depends();
/* When the buffer fills up discard new entries. Assume /* When the buffer fills up discard new entries. Assume
that the earlier errors are the more interesting. */ that the earlier errors are the more interesting. */
if (entry >= MCE_LOG_LEN) { if (entry >= MCE_LOG_LEN) {
...@@ -333,8 +332,7 @@ static ssize_t mce_read(struct file *filp, char __user *ubuf, size_t usize, loff ...@@ -333,8 +332,7 @@ static ssize_t mce_read(struct file *filp, char __user *ubuf, size_t usize, loff
int i, err; int i, err;
down(&mce_read_sem); down(&mce_read_sem);
next = mcelog.next; next = rcu_dereference(mcelog.next);
read_barrier_depends();
/* Only supports full reads right now */ /* Only supports full reads right now */
if (*off != 0 || usize < MCE_LOG_LEN*sizeof(struct mce)) { if (*off != 0 || usize < MCE_LOG_LEN*sizeof(struct mce)) {
......
...@@ -613,7 +613,7 @@ void shrink_dcache_parent(struct dentry * parent) ...@@ -613,7 +613,7 @@ void shrink_dcache_parent(struct dentry * parent)
* *
* Prune the dentries that are anonymous * Prune the dentries that are anonymous
* *
* parsing d_hash list does not read_barrier_depends() as it * parsing d_hash list does not hlist_for_each_rcu() as it
* done under dcache_lock. * done under dcache_lock.
* *
*/ */
...@@ -970,11 +970,10 @@ struct dentry * __d_lookup(struct dentry * parent, struct qstr * name) ...@@ -970,11 +970,10 @@ struct dentry * __d_lookup(struct dentry * parent, struct qstr * name)
rcu_read_lock(); rcu_read_lock();
hlist_for_each (node, head) { hlist_for_each_rcu(node, head) {
struct dentry *dentry; struct dentry *dentry;
struct qstr *qstr; struct qstr *qstr;
smp_read_barrier_depends();
dentry = hlist_entry(node, struct dentry, d_hash); dentry = hlist_entry(node, struct dentry, d_hash);
smp_rmb(); smp_rmb();
...@@ -1001,8 +1000,7 @@ struct dentry * __d_lookup(struct dentry * parent, struct qstr * name) ...@@ -1001,8 +1000,7 @@ struct dentry * __d_lookup(struct dentry * parent, struct qstr * name)
if (dentry->d_parent != parent) if (dentry->d_parent != parent)
goto next; goto next;
qstr = &dentry->d_name; qstr = rcu_dereference(&dentry->d_name);
smp_read_barrier_depends();
if (parent->d_op && parent->d_op->d_compare) { if (parent->d_op && parent->d_op->d_compare) {
if (parent->d_op->d_compare(parent, qstr, name)) if (parent->d_op->d_compare(parent, qstr, name))
goto next; goto next;
...@@ -1055,7 +1053,7 @@ int d_validate(struct dentry *dentry, struct dentry *dparent) ...@@ -1055,7 +1053,7 @@ int d_validate(struct dentry *dentry, struct dentry *dparent)
spin_lock(&dcache_lock); spin_lock(&dcache_lock);
base = d_hash(dparent, dentry->d_name.hash); base = d_hash(dparent, dentry->d_name.hash);
hlist_for_each(lhp,base) { hlist_for_each(lhp,base) {
/* read_barrier_depends() not required for d_hash list /* hlist_for_each_rcu() not required for d_hash list
* as it is parsed under dcache_lock * as it is parsed under dcache_lock
*/ */
if (dentry == hlist_entry(lhp, struct dentry, d_hash)) { if (dentry == hlist_entry(lhp, struct dentry, d_hash)) {
......
...@@ -420,11 +420,11 @@ static inline void list_splice_init(struct list_head *list, ...@@ -420,11 +420,11 @@ static inline void list_splice_init(struct list_head *list,
*/ */
#define list_for_each_rcu(pos, head) \ #define list_for_each_rcu(pos, head) \
for (pos = (head)->next, prefetch(pos->next); pos != (head); \ for (pos = (head)->next, prefetch(pos->next); pos != (head); \
pos = pos->next, ({ smp_read_barrier_depends(); 0;}), prefetch(pos->next)) pos = rcu_dereference(pos->next), prefetch(pos->next))
#define __list_for_each_rcu(pos, head) \ #define __list_for_each_rcu(pos, head) \
for (pos = (head)->next; pos != (head); \ for (pos = (head)->next; pos != (head); \
pos = pos->next, ({ smp_read_barrier_depends(); 0;})) pos = rcu_dereference(pos->next))
/** /**
* list_for_each_safe_rcu - iterate over an rcu-protected list safe * list_for_each_safe_rcu - iterate over an rcu-protected list safe
...@@ -439,7 +439,7 @@ static inline void list_splice_init(struct list_head *list, ...@@ -439,7 +439,7 @@ static inline void list_splice_init(struct list_head *list,
*/ */
#define list_for_each_safe_rcu(pos, n, head) \ #define list_for_each_safe_rcu(pos, n, head) \
for (pos = (head)->next, n = pos->next; pos != (head); \ for (pos = (head)->next, n = pos->next; pos != (head); \
pos = n, ({ smp_read_barrier_depends(); 0;}), n = pos->next) pos = rcu_dereference(n), n = pos->next)
/** /**
* list_for_each_entry_rcu - iterate over rcu list of given type * list_for_each_entry_rcu - iterate over rcu list of given type
...@@ -455,8 +455,8 @@ static inline void list_splice_init(struct list_head *list, ...@@ -455,8 +455,8 @@ static inline void list_splice_init(struct list_head *list,
for (pos = list_entry((head)->next, typeof(*pos), member), \ for (pos = list_entry((head)->next, typeof(*pos), member), \
prefetch(pos->member.next); \ prefetch(pos->member.next); \
&pos->member != (head); \ &pos->member != (head); \
pos = list_entry(pos->member.next, typeof(*pos), member), \ pos = rcu_dereference(list_entry(pos->member.next, \
({ smp_read_barrier_depends(); 0;}), \ typeof(*pos), member)), \
prefetch(pos->member.next)) prefetch(pos->member.next))
...@@ -472,7 +472,7 @@ static inline void list_splice_init(struct list_head *list, ...@@ -472,7 +472,7 @@ static inline void list_splice_init(struct list_head *list,
*/ */
#define list_for_each_continue_rcu(pos, head) \ #define list_for_each_continue_rcu(pos, head) \
for ((pos) = (pos)->next, prefetch((pos)->next); (pos) != (head); \ for ((pos) = (pos)->next, prefetch((pos)->next); (pos) != (head); \
(pos) = (pos)->next, ({ smp_read_barrier_depends(); 0;}), prefetch((pos)->next)) (pos) = rcu_dereference((pos)->next), prefetch((pos)->next))
/* /*
* Double linked lists with a single pointer list head. * Double linked lists with a single pointer list head.
...@@ -578,12 +578,9 @@ static inline void hlist_add_head(struct hlist_node *n, struct hlist_head *h) ...@@ -578,12 +578,9 @@ static inline void hlist_add_head(struct hlist_node *n, struct hlist_head *h)
* or hlist_del_rcu(), running on this same list. * or hlist_del_rcu(), running on this same list.
* However, it is perfectly legal to run concurrently with * However, it is perfectly legal to run concurrently with
* the _rcu list-traversal primitives, such as * the _rcu list-traversal primitives, such as
* hlist_for_each_entry(), but only if smp_read_barrier_depends() * hlist_for_each_rcu(), used to prevent memory-consistency
* is used to prevent memory-consistency problems on Alpha CPUs. * problems on Alpha CPUs. Regardless of the type of CPU, the
* Regardless of the type of CPU, the list-traversal primitive * list-traversal primitive must be guarded by rcu_read_lock().
* must be guarded by rcu_read_lock().
*
* OK, so why don't we have an hlist_for_each_entry_rcu()???
*/ */
static inline void hlist_add_head_rcu(struct hlist_node *n, static inline void hlist_add_head_rcu(struct hlist_node *n,
struct hlist_head *h) struct hlist_head *h)
...@@ -628,6 +625,10 @@ static inline void hlist_add_after(struct hlist_node *n, ...@@ -628,6 +625,10 @@ static inline void hlist_add_after(struct hlist_node *n,
for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \ for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
pos = n) pos = n)
#define hlist_for_each_rcu(pos, head) \
for ((pos) = (head)->first; pos && ({ prefetch((pos)->next); 1; }); \
(pos) = rcu_dereference((pos)->next))
/** /**
* hlist_for_each_entry - iterate over list of given type * hlist_for_each_entry - iterate over list of given type
* @tpos: the type * to use as a loop counter. * @tpos: the type * to use as a loop counter.
...@@ -693,7 +694,7 @@ static inline void hlist_add_after(struct hlist_node *n, ...@@ -693,7 +694,7 @@ static inline void hlist_add_after(struct hlist_node *n,
for (pos = (head)->first; \ for (pos = (head)->first; \
pos && ({ prefetch(pos->next); 1;}) && \ pos && ({ prefetch(pos->next); 1;}) && \
({ tpos = hlist_entry(pos, typeof(*tpos), member); 1;}); \ ({ tpos = hlist_entry(pos, typeof(*tpos), member); 1;}); \
pos = pos->next, ({ smp_read_barrier_depends(); 0; }) ) pos = rcu_dereference(pos->next))
#else #else
#warning "don't include kernel headers in userspace" #warning "don't include kernel headers in userspace"
......
...@@ -222,6 +222,22 @@ static inline int rcu_pending(int cpu) ...@@ -222,6 +222,22 @@ static inline int rcu_pending(int cpu)
*/ */
#define rcu_read_unlock_bh() local_bh_enable() #define rcu_read_unlock_bh() local_bh_enable()
/**
* rcu_dereference - fetch an RCU-protected pointer in an
* RCU read-side critical section. This pointer may later
* be safely dereferenced.
*
* Inserts memory barriers on architectures that require them
* (currently only the Alpha), and, more importantly, documents
* exactly which pointers are protected by RCU.
*/
#define rcu_dereference(p) ({ \
typeof(p) _________p1 = p; \
smp_read_barrier_depends(); \
(_________p1); \
})
extern void rcu_init(void); extern void rcu_init(void);
extern void rcu_check_callbacks(int cpu, int user); extern void rcu_check_callbacks(int cpu, int user);
extern void rcu_restart_cpu(int cpu); extern void rcu_restart_cpu(int cpu);
......
...@@ -100,7 +100,7 @@ int ipc_findkey(struct ipc_ids* ids, key_t key) ...@@ -100,7 +100,7 @@ int ipc_findkey(struct ipc_ids* ids, key_t key)
int max_id = ids->max_id; int max_id = ids->max_id;
/* /*
* read_barrier_depends is not needed here * rcu_dereference() is not needed here
* since ipc_ids.sem is held * since ipc_ids.sem is held
*/ */
for (id = 0; id <= max_id; id++) { for (id = 0; id <= max_id; id++) {
...@@ -171,7 +171,7 @@ int ipc_addid(struct ipc_ids* ids, struct kern_ipc_perm* new, int size) ...@@ -171,7 +171,7 @@ int ipc_addid(struct ipc_ids* ids, struct kern_ipc_perm* new, int size)
size = grow_ary(ids,size); size = grow_ary(ids,size);
/* /*
* read_barrier_depends() is not needed here since * rcu_dereference()() is not needed here since
* ipc_ids.sem is held * ipc_ids.sem is held
*/ */
for (id = 0; id < size; id++) { for (id = 0; id < size; id++) {
...@@ -220,7 +220,7 @@ struct kern_ipc_perm* ipc_rmid(struct ipc_ids* ids, int id) ...@@ -220,7 +220,7 @@ struct kern_ipc_perm* ipc_rmid(struct ipc_ids* ids, int id)
BUG(); BUG();
/* /*
* do not need a read_barrier_depends() here to force ordering * do not need a rcu_dereference()() here to force ordering
* on Alpha, since the ipc_ids.sem is held. * on Alpha, since the ipc_ids.sem is held.
*/ */
p = ids->entries[lid].p; p = ids->entries[lid].p;
...@@ -515,13 +515,12 @@ struct kern_ipc_perm* ipc_lock(struct ipc_ids* ids, int id) ...@@ -515,13 +515,12 @@ struct kern_ipc_perm* ipc_lock(struct ipc_ids* ids, int id)
* Note: The following two read barriers are corresponding * Note: The following two read barriers are corresponding
* to the two write barriers in grow_ary(). They guarantee * to the two write barriers in grow_ary(). They guarantee
* the writes are seen in the same order on the read side. * the writes are seen in the same order on the read side.
* smp_rmb() has effect on all CPUs. read_barrier_depends() * smp_rmb() has effect on all CPUs. rcu_dereference()
* is used if there are data dependency between two reads, and * is used if there are data dependency between two reads, and
* has effect only on Alpha. * has effect only on Alpha.
*/ */
smp_rmb(); /* prevent indexing old array with new size */ smp_rmb(); /* prevent indexing old array with new size */
entries = ids->entries; entries = rcu_dereference(ids->entries);
read_barrier_depends(); /*prevent seeing new array unitialized */
out = entries[lid].p; out = entries[lid].p;
if(out == NULL) { if(out == NULL) {
rcu_read_unlock(); rcu_read_unlock();
......
...@@ -56,8 +56,7 @@ int br_handle_frame_finish(struct sk_buff *skb) ...@@ -56,8 +56,7 @@ int br_handle_frame_finish(struct sk_buff *skb)
dest = skb->mac.ethernet->h_dest; dest = skb->mac.ethernet->h_dest;
rcu_read_lock(); rcu_read_lock();
p = skb->dev->br_port; p = rcu_dereference(skb->dev->br_port);
smp_read_barrier_depends();
if (p == NULL || p->state == BR_STATE_DISABLED) { if (p == NULL || p->state == BR_STATE_DISABLED) {
kfree_skb(skb); kfree_skb(skb);
......
...@@ -1332,8 +1332,7 @@ int dev_queue_xmit(struct sk_buff *skb) ...@@ -1332,8 +1332,7 @@ int dev_queue_xmit(struct sk_buff *skb)
* also serializes access to the device queue. * also serializes access to the device queue.
*/ */
q = dev->qdisc; q = rcu_dereference(dev->qdisc);
smp_read_barrier_depends();
#ifdef CONFIG_NET_CLS_ACT #ifdef CONFIG_NET_CLS_ACT
skb->tc_verd = SET_TC_AT(skb->tc_verd,AT_EGRESS); skb->tc_verd = SET_TC_AT(skb->tc_verd,AT_EGRESS);
#endif #endif
......
...@@ -783,13 +783,12 @@ void nf_log_packet(int pf, ...@@ -783,13 +783,12 @@ void nf_log_packet(int pf,
nf_logfn *logfn; nf_logfn *logfn;
rcu_read_lock(); rcu_read_lock();
logfn = nf_logging[pf]; logfn = rcu_dereference(nf_logging[pf]);
if (logfn) { if (logfn) {
va_start(args, fmt); va_start(args, fmt);
vsnprintf(prefix, sizeof(prefix), fmt, args); vsnprintf(prefix, sizeof(prefix), fmt, args);
va_end(args); va_end(args);
/* We must read logging before nf_logfn[pf] */ /* We must read logging before nf_logfn[pf] */
smp_read_barrier_depends();
logfn(hooknum, skb, in, out, prefix); logfn(hooknum, skb, in, out, prefix);
} else if (!reported) { } else if (!reported) {
printk(KERN_WARNING "nf_log_packet: can\'t log yet, " printk(KERN_WARNING "nf_log_packet: can\'t log yet, "
......
/* /*
* DECnet An implementation of the DECnet protocol suite for the LINUX * DECnet An implementation of the DECnet protocol suite for the LINUX
* operating system. DECnet is implemented using the BSD Socket * operating system. DECnet is implemented using the BSD Socket
...@@ -1175,8 +1174,8 @@ static int __dn_route_output_key(struct dst_entry **pprt, const struct flowi *fl ...@@ -1175,8 +1174,8 @@ static int __dn_route_output_key(struct dst_entry **pprt, const struct flowi *fl
if (!(flags & MSG_TRYHARD)) { if (!(flags & MSG_TRYHARD)) {
rcu_read_lock_bh(); rcu_read_lock_bh();
for(rt = dn_rt_hash_table[hash].chain; rt; rt = rt->u.rt_next) { for(rt = rcu_dereference(dn_rt_hash_table[hash].chain); rt;
smp_read_barrier_depends(); rt = rcu_dereference(rt->u.rt_next)) {
if ((flp->fld_dst == rt->fl.fld_dst) && if ((flp->fld_dst == rt->fl.fld_dst) &&
(flp->fld_src == rt->fl.fld_src) && (flp->fld_src == rt->fl.fld_src) &&
#ifdef CONFIG_DECNET_ROUTE_FWMARK #ifdef CONFIG_DECNET_ROUTE_FWMARK
...@@ -1454,8 +1453,8 @@ int dn_route_input(struct sk_buff *skb) ...@@ -1454,8 +1453,8 @@ int dn_route_input(struct sk_buff *skb)
return 0; return 0;
rcu_read_lock(); rcu_read_lock();
for(rt = dn_rt_hash_table[hash].chain; rt != NULL; rt = rt->u.rt_next) { for(rt = rcu_dereference(dn_rt_hash_table[hash].chain); rt != NULL;
read_barrier_depends(); rt = rcu_dereference(rt->u.rt_next)) {
if ((rt->fl.fld_src == cb->src) && if ((rt->fl.fld_src == cb->src) &&
(rt->fl.fld_dst == cb->dst) && (rt->fl.fld_dst == cb->dst) &&
(rt->fl.oif == 0) && (rt->fl.oif == 0) &&
...@@ -1648,8 +1647,9 @@ int dn_cache_dump(struct sk_buff *skb, struct netlink_callback *cb) ...@@ -1648,8 +1647,9 @@ int dn_cache_dump(struct sk_buff *skb, struct netlink_callback *cb)
if (h > s_h) if (h > s_h)
s_idx = 0; s_idx = 0;
rcu_read_lock_bh(); rcu_read_lock_bh();
for(rt = dn_rt_hash_table[h].chain, idx = 0; rt; rt = rt->u.rt_next, idx++) { for(rt = rcu_dereference(dn_rt_hash_table[h].chain), idx = 0;
smp_read_barrier_depends(); rt;
rt = rcu_dereference(rt->u.rt_next), idx++) {
if (idx < s_idx) if (idx < s_idx)
continue; continue;
skb->dst = dst_clone(&rt->u.dst); skb->dst = dst_clone(&rt->u.dst);
...@@ -1692,9 +1692,8 @@ static struct dn_route *dn_rt_cache_get_first(struct seq_file *seq) ...@@ -1692,9 +1692,8 @@ static struct dn_route *dn_rt_cache_get_first(struct seq_file *seq)
static struct dn_route *dn_rt_cache_get_next(struct seq_file *seq, struct dn_route *rt) static struct dn_route *dn_rt_cache_get_next(struct seq_file *seq, struct dn_route *rt)
{ {
struct dn_rt_cache_iter_state *s = seq->private; struct dn_rt_cache_iter_state *s = rcu_dereference(seq->private);
smp_read_barrier_depends();
rt = rt->u.rt_next; rt = rt->u.rt_next;
while(!rt) { while(!rt) {
rcu_read_unlock_bh(); rcu_read_unlock_bh();
......
...@@ -705,8 +705,7 @@ static void icmp_unreach(struct sk_buff *skb) ...@@ -705,8 +705,7 @@ static void icmp_unreach(struct sk_buff *skb)
read_unlock(&raw_v4_lock); read_unlock(&raw_v4_lock);
rcu_read_lock(); rcu_read_lock();
ipprot = inet_protos[hash]; ipprot = rcu_dereference(inet_protos[hash]);
smp_read_barrier_depends();
if (ipprot && ipprot->err_handler) if (ipprot && ipprot->err_handler)
ipprot->err_handler(skb, info); ipprot->err_handler(skb, info);
rcu_read_unlock(); rcu_read_unlock();
......
...@@ -231,10 +231,9 @@ static inline int ip_local_deliver_finish(struct sk_buff *skb) ...@@ -231,10 +231,9 @@ static inline int ip_local_deliver_finish(struct sk_buff *skb)
if (raw_sk) if (raw_sk)
raw_v4_input(skb, skb->nh.iph, hash); raw_v4_input(skb, skb->nh.iph, hash);
if ((ipprot = inet_protos[hash]) != NULL) { if ((ipprot = rcu_dereference(inet_protos[hash])) != NULL) {
int ret; int ret;
smp_read_barrier_depends();
if (!ipprot->no_policy && if (!ipprot->no_policy &&
!xfrm4_policy_check(NULL, XFRM_POLICY_IN, skb)) { !xfrm4_policy_check(NULL, XFRM_POLICY_IN, skb)) {
kfree_skb(skb); kfree_skb(skb);
......
...@@ -237,9 +237,8 @@ static struct rtable *rt_cache_get_first(struct seq_file *seq) ...@@ -237,9 +237,8 @@ static struct rtable *rt_cache_get_first(struct seq_file *seq)
static struct rtable *rt_cache_get_next(struct seq_file *seq, struct rtable *r) static struct rtable *rt_cache_get_next(struct seq_file *seq, struct rtable *r)
{ {
struct rt_cache_iter_state *st = seq->private; struct rt_cache_iter_state *st = rcu_dereference(seq->private);
smp_read_barrier_depends();
r = r->u.rt_next; r = r->u.rt_next;
while (!r) { while (!r) {
rcu_read_unlock_bh(); rcu_read_unlock_bh();
...@@ -1004,10 +1003,9 @@ void ip_rt_redirect(u32 old_gw, u32 daddr, u32 new_gw, ...@@ -1004,10 +1003,9 @@ void ip_rt_redirect(u32 old_gw, u32 daddr, u32 new_gw,
rthp=&rt_hash_table[hash].chain; rthp=&rt_hash_table[hash].chain;
rcu_read_lock(); rcu_read_lock();
while ((rth = *rthp) != NULL) { while ((rth = rcu_dereference(*rthp)) != NULL) {
struct rtable *rt; struct rtable *rt;
smp_read_barrier_depends();
if (rth->fl.fl4_dst != daddr || if (rth->fl.fl4_dst != daddr ||
rth->fl.fl4_src != skeys[i] || rth->fl.fl4_src != skeys[i] ||
rth->fl.fl4_tos != tos || rth->fl.fl4_tos != tos ||
...@@ -1259,9 +1257,8 @@ unsigned short ip_rt_frag_needed(struct iphdr *iph, unsigned short new_mtu) ...@@ -1259,9 +1257,8 @@ unsigned short ip_rt_frag_needed(struct iphdr *iph, unsigned short new_mtu)
unsigned hash = rt_hash_code(daddr, skeys[i], tos); unsigned hash = rt_hash_code(daddr, skeys[i], tos);
rcu_read_lock(); rcu_read_lock();
for (rth = rt_hash_table[hash].chain; rth; for (rth = rcu_dereference(rt_hash_table[hash].chain); rth;
rth = rth->u.rt_next) { rth = rcu_dereference(rth->u.rt_next)) {
smp_read_barrier_depends();
if (rth->fl.fl4_dst == daddr && if (rth->fl.fl4_dst == daddr &&
rth->fl.fl4_src == skeys[i] && rth->fl.fl4_src == skeys[i] &&
rth->rt_dst == daddr && rth->rt_dst == daddr &&
...@@ -1864,8 +1861,8 @@ int ip_route_input(struct sk_buff *skb, u32 daddr, u32 saddr, ...@@ -1864,8 +1861,8 @@ int ip_route_input(struct sk_buff *skb, u32 daddr, u32 saddr,
hash = rt_hash_code(daddr, saddr ^ (iif << 5), tos); hash = rt_hash_code(daddr, saddr ^ (iif << 5), tos);
rcu_read_lock(); rcu_read_lock();
for (rth = rt_hash_table[hash].chain; rth; rth = rth->u.rt_next) { for (rth = rcu_dereference(rt_hash_table[hash].chain); rth;
smp_read_barrier_depends(); rth = rcu_dereference(rth->u.rt_next)) {
if (rth->fl.fl4_dst == daddr && if (rth->fl.fl4_dst == daddr &&
rth->fl.fl4_src == saddr && rth->fl.fl4_src == saddr &&
rth->fl.iif == iif && rth->fl.iif == iif &&
...@@ -2232,8 +2229,8 @@ int __ip_route_output_key(struct rtable **rp, const struct flowi *flp) ...@@ -2232,8 +2229,8 @@ int __ip_route_output_key(struct rtable **rp, const struct flowi *flp)
hash = rt_hash_code(flp->fl4_dst, flp->fl4_src ^ (flp->oif << 5), flp->fl4_tos); hash = rt_hash_code(flp->fl4_dst, flp->fl4_src ^ (flp->oif << 5), flp->fl4_tos);
rcu_read_lock_bh(); rcu_read_lock_bh();
for (rth = rt_hash_table[hash].chain; rth; rth = rth->u.rt_next) { for (rth = rcu_dereference(rt_hash_table[hash].chain); rth;
smp_read_barrier_depends(); rth = rcu_dereference(rth->u.rt_next)) {
if (rth->fl.fl4_dst == flp->fl4_dst && if (rth->fl.fl4_dst == flp->fl4_dst &&
rth->fl.fl4_src == flp->fl4_src && rth->fl.fl4_src == flp->fl4_src &&
rth->fl.iif == 0 && rth->fl.iif == 0 &&
...@@ -2464,9 +2461,8 @@ int ip_rt_dump(struct sk_buff *skb, struct netlink_callback *cb) ...@@ -2464,9 +2461,8 @@ int ip_rt_dump(struct sk_buff *skb, struct netlink_callback *cb)
if (h > s_h) if (h > s_h)
s_idx = 0; s_idx = 0;
rcu_read_lock_bh(); rcu_read_lock_bh();
for (rt = rt_hash_table[h].chain, idx = 0; rt; for (rt = rcu_dereference(rt_hash_table[h].chain), idx = 0; rt;
rt = rt->u.rt_next, idx++) { rt = rcu_dereference(rt->u.rt_next), idx++) {
smp_read_barrier_depends();
if (idx < s_idx) if (idx < s_idx)
continue; continue;
skb->dst = dst_clone(&rt->u.dst); skb->dst = dst_clone(&rt->u.dst);
......
...@@ -530,8 +530,7 @@ static void icmpv6_notify(struct sk_buff *skb, int type, int code, u32 info) ...@@ -530,8 +530,7 @@ static void icmpv6_notify(struct sk_buff *skb, int type, int code, u32 info)
hash = nexthdr & (MAX_INET_PROTOS - 1); hash = nexthdr & (MAX_INET_PROTOS - 1);
rcu_read_lock(); rcu_read_lock();
ipprot = inet6_protos[hash]; ipprot = rcu_dereference(inet6_protos[hash]);
smp_read_barrier_depends();
if (ipprot && ipprot->err_handler) if (ipprot && ipprot->err_handler)
ipprot->err_handler(skb, NULL, type, code, inner_offset, info); ipprot->err_handler(skb, NULL, type, code, inner_offset, info);
rcu_read_unlock(); rcu_read_unlock();
......
...@@ -167,10 +167,9 @@ static inline int ip6_input_finish(struct sk_buff *skb) ...@@ -167,10 +167,9 @@ static inline int ip6_input_finish(struct sk_buff *skb)
ipv6_raw_deliver(skb, nexthdr); ipv6_raw_deliver(skb, nexthdr);
hash = nexthdr & (MAX_INET_PROTOS - 1); hash = nexthdr & (MAX_INET_PROTOS - 1);
if ((ipprot = inet6_protos[hash]) != NULL) { if ((ipprot = rcu_dereference(inet6_protos[hash])) != NULL) {
int ret; int ret;
smp_read_barrier_depends();
if (ipprot->flags & INET6_PROTO_FINAL) { if (ipprot->flags & INET6_PROTO_FINAL) {
struct ipv6hdr *hdr; struct ipv6hdr *hdr;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment