Commit 702b71e7 authored by Jozsef Kadlecsik's avatar Jozsef Kadlecsik

netfilter: ipset: Add element count to all set types header

It is better to list the set elements for all set types, thus the
header information is uniform. Element counts are therefore added
to the bitmap and list types.
Signed-off-by: default avatarJozsef Kadlecsik <kadlec@blackhole.kfki.hu>
parent a54dad51
...@@ -250,6 +250,8 @@ struct ip_set { ...@@ -250,6 +250,8 @@ struct ip_set {
u8 flags; u8 flags;
/* Default timeout value, if enabled */ /* Default timeout value, if enabled */
u32 timeout; u32 timeout;
/* Number of elements (vs timeout) */
u32 elements;
/* Element data size */ /* Element data size */
size_t dsize; size_t dsize;
/* Offsets to extensions in elements */ /* Offsets to extensions in elements */
......
...@@ -6,8 +6,8 @@ ...@@ -6,8 +6,8 @@
#define IPSET_BITMAP_MAX_RANGE 0x0000FFFF #define IPSET_BITMAP_MAX_RANGE 0x0000FFFF
enum { enum {
IPSET_ADD_STORE_PLAIN_TIMEOUT = -1,
IPSET_ADD_FAILED = 1, IPSET_ADD_FAILED = 1,
IPSET_ADD_STORE_PLAIN_TIMEOUT,
IPSET_ADD_START_STORED_TIMEOUT, IPSET_ADD_START_STORED_TIMEOUT,
}; };
......
...@@ -83,6 +83,7 @@ mtype_flush(struct ip_set *set) ...@@ -83,6 +83,7 @@ mtype_flush(struct ip_set *set)
if (set->extensions & IPSET_EXT_DESTROY) if (set->extensions & IPSET_EXT_DESTROY)
mtype_ext_cleanup(set); mtype_ext_cleanup(set);
memset(map->members, 0, map->memsize); memset(map->members, 0, map->memsize);
set->elements = 0;
} }
/* Calculate the actual memory size of the set data */ /* Calculate the actual memory size of the set data */
...@@ -105,7 +106,8 @@ mtype_head(struct ip_set *set, struct sk_buff *skb) ...@@ -105,7 +106,8 @@ mtype_head(struct ip_set *set, struct sk_buff *skb)
goto nla_put_failure; goto nla_put_failure;
if (mtype_do_head(skb, map) || if (mtype_do_head(skb, map) ||
nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) || nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) ||
nla_put_net32(skb, IPSET_ATTR_MEMSIZE, htonl(memsize))) nla_put_net32(skb, IPSET_ATTR_MEMSIZE, htonl(memsize)) ||
nla_put_net32(skb, IPSET_ATTR_ELEMENTS, htonl(set->elements)))
goto nla_put_failure; goto nla_put_failure;
if (unlikely(ip_set_put_flags(skb, set))) if (unlikely(ip_set_put_flags(skb, set)))
goto nla_put_failure; goto nla_put_failure;
...@@ -149,6 +151,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -149,6 +151,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
if (ret == IPSET_ADD_FAILED) { if (ret == IPSET_ADD_FAILED) {
if (SET_WITH_TIMEOUT(set) && if (SET_WITH_TIMEOUT(set) &&
ip_set_timeout_expired(ext_timeout(x, set))) { ip_set_timeout_expired(ext_timeout(x, set))) {
set->elements--;
ret = 0; ret = 0;
} else if (!(flags & IPSET_FLAG_EXIST)) { } else if (!(flags & IPSET_FLAG_EXIST)) {
set_bit(e->id, map->members); set_bit(e->id, map->members);
...@@ -157,6 +160,8 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -157,6 +160,8 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
/* Element is re-added, cleanup extensions */ /* Element is re-added, cleanup extensions */
ip_set_ext_destroy(set, x); ip_set_ext_destroy(set, x);
} }
if (ret > 0)
set->elements--;
if (SET_WITH_TIMEOUT(set)) if (SET_WITH_TIMEOUT(set))
#ifdef IP_SET_BITMAP_STORED_TIMEOUT #ifdef IP_SET_BITMAP_STORED_TIMEOUT
...@@ -174,6 +179,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -174,6 +179,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
/* Activate element */ /* Activate element */
set_bit(e->id, map->members); set_bit(e->id, map->members);
set->elements++;
return 0; return 0;
} }
...@@ -190,6 +196,7 @@ mtype_del(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -190,6 +196,7 @@ mtype_del(struct ip_set *set, void *value, const struct ip_set_ext *ext,
return -IPSET_ERR_EXIST; return -IPSET_ERR_EXIST;
ip_set_ext_destroy(set, x); ip_set_ext_destroy(set, x);
set->elements--;
if (SET_WITH_TIMEOUT(set) && if (SET_WITH_TIMEOUT(set) &&
ip_set_timeout_expired(ext_timeout(x, set))) ip_set_timeout_expired(ext_timeout(x, set)))
return -IPSET_ERR_EXIST; return -IPSET_ERR_EXIST;
...@@ -285,6 +292,7 @@ mtype_gc(unsigned long ul_set) ...@@ -285,6 +292,7 @@ mtype_gc(unsigned long ul_set)
if (ip_set_timeout_expired(ext_timeout(x, set))) { if (ip_set_timeout_expired(ext_timeout(x, set))) {
clear_bit(id, map->members); clear_bit(id, map->members);
ip_set_ext_destroy(set, x); ip_set_ext_destroy(set, x);
set->elements--;
} }
} }
spin_unlock_bh(&set->lock); spin_unlock_bh(&set->lock);
......
...@@ -275,7 +275,6 @@ htable_bits(u32 hashsize) ...@@ -275,7 +275,6 @@ htable_bits(u32 hashsize)
struct htype { struct htype {
struct htable __rcu *table; /* the hash table */ struct htable __rcu *table; /* the hash table */
u32 maxelem; /* max elements in the hash */ u32 maxelem; /* max elements in the hash */
u32 elements; /* current element (vs timeout) */
u32 initval; /* random jhash init value */ u32 initval; /* random jhash init value */
#ifdef IP_SET_HASH_WITH_MARKMASK #ifdef IP_SET_HASH_WITH_MARKMASK
u32 markmask; /* markmask value for mark mask to store */ u32 markmask; /* markmask value for mark mask to store */
...@@ -400,7 +399,7 @@ mtype_flush(struct ip_set *set) ...@@ -400,7 +399,7 @@ mtype_flush(struct ip_set *set)
#ifdef IP_SET_HASH_WITH_NETS #ifdef IP_SET_HASH_WITH_NETS
memset(h->nets, 0, sizeof(struct net_prefixes) * NLEN(set->family)); memset(h->nets, 0, sizeof(struct net_prefixes) * NLEN(set->family));
#endif #endif
h->elements = 0; set->elements = 0;
} }
/* Destroy the hashtable part of the set */ /* Destroy the hashtable part of the set */
...@@ -506,7 +505,7 @@ mtype_expire(struct ip_set *set, struct htype *h, u8 nets_length, size_t dsize) ...@@ -506,7 +505,7 @@ mtype_expire(struct ip_set *set, struct htype *h, u8 nets_length, size_t dsize)
nets_length, k); nets_length, k);
#endif #endif
ip_set_ext_destroy(set, data); ip_set_ext_destroy(set, data);
h->elements--; set->elements--;
d++; d++;
} }
} }
...@@ -715,11 +714,11 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -715,11 +714,11 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
bool deleted = false, forceadd = false, reuse = false; bool deleted = false, forceadd = false, reuse = false;
u32 key, multi = 0; u32 key, multi = 0;
if (h->elements >= h->maxelem) { if (set->elements >= h->maxelem) {
if (SET_WITH_TIMEOUT(set)) if (SET_WITH_TIMEOUT(set))
/* FIXME: when set is full, we slow down here */ /* FIXME: when set is full, we slow down here */
mtype_expire(set, h, NLEN(set->family), set->dsize); mtype_expire(set, h, NLEN(set->family), set->dsize);
if (h->elements >= h->maxelem && SET_WITH_FORCEADD(set)) if (set->elements >= h->maxelem && SET_WITH_FORCEADD(set))
forceadd = true; forceadd = true;
} }
...@@ -732,7 +731,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -732,7 +731,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
pr_warn("Set %s is full, maxelem %u reached\n", pr_warn("Set %s is full, maxelem %u reached\n",
set->name, h->maxelem); set->name, h->maxelem);
return -IPSET_ERR_HASH_FULL; return -IPSET_ERR_HASH_FULL;
} else if (h->elements >= h->maxelem) { } else if (set->elements >= h->maxelem) {
goto set_full; goto set_full;
} }
old = NULL; old = NULL;
...@@ -781,11 +780,11 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -781,11 +780,11 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
NLEN(set->family), i); NLEN(set->family), i);
#endif #endif
ip_set_ext_destroy(set, data); ip_set_ext_destroy(set, data);
h->elements--; set->elements--;
} }
goto copy_data; goto copy_data;
} }
if (h->elements >= h->maxelem) if (set->elements >= h->maxelem)
goto set_full; goto set_full;
/* Create a new slot */ /* Create a new slot */
if (n->pos >= n->size) { if (n->pos >= n->size) {
...@@ -810,7 +809,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -810,7 +809,7 @@ mtype_add(struct ip_set *set, void *value, const struct ip_set_ext *ext,
j = n->pos++; j = n->pos++;
data = ahash_data(n, j, set->dsize); data = ahash_data(n, j, set->dsize);
copy_data: copy_data:
h->elements++; set->elements++;
#ifdef IP_SET_HASH_WITH_NETS #ifdef IP_SET_HASH_WITH_NETS
for (i = 0; i < IPSET_NET_COUNT; i++) for (i = 0; i < IPSET_NET_COUNT; i++)
mtype_add_cidr(h, NCIDR_PUT(DCIDR_GET(d->cidr, i)), mtype_add_cidr(h, NCIDR_PUT(DCIDR_GET(d->cidr, i)),
...@@ -883,7 +882,7 @@ mtype_del(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -883,7 +882,7 @@ mtype_del(struct ip_set *set, void *value, const struct ip_set_ext *ext,
smp_mb__after_atomic(); smp_mb__after_atomic();
if (i + 1 == n->pos) if (i + 1 == n->pos)
n->pos--; n->pos--;
h->elements--; set->elements--;
#ifdef IP_SET_HASH_WITH_NETS #ifdef IP_SET_HASH_WITH_NETS
for (j = 0; j < IPSET_NET_COUNT; j++) for (j = 0; j < IPSET_NET_COUNT; j++)
mtype_del_cidr(h, NCIDR_PUT(DCIDR_GET(d->cidr, j)), mtype_del_cidr(h, NCIDR_PUT(DCIDR_GET(d->cidr, j)),
...@@ -1084,7 +1083,7 @@ mtype_head(struct ip_set *set, struct sk_buff *skb) ...@@ -1084,7 +1083,7 @@ mtype_head(struct ip_set *set, struct sk_buff *skb)
#endif #endif
if (nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) || if (nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) ||
nla_put_net32(skb, IPSET_ATTR_MEMSIZE, htonl(memsize)) || nla_put_net32(skb, IPSET_ATTR_MEMSIZE, htonl(memsize)) ||
nla_put_net32(skb, IPSET_ATTR_ELEMENTS, htonl(h->elements))) nla_put_net32(skb, IPSET_ATTR_ELEMENTS, htonl(set->elements)))
goto nla_put_failure; goto nla_put_failure;
if (unlikely(ip_set_put_flags(skb, set))) if (unlikely(ip_set_put_flags(skb, set)))
goto nla_put_failure; goto nla_put_failure;
......
...@@ -166,6 +166,7 @@ __list_set_del_rcu(struct rcu_head * rcu) ...@@ -166,6 +166,7 @@ __list_set_del_rcu(struct rcu_head * rcu)
static inline void static inline void
list_set_del(struct ip_set *set, struct set_elem *e) list_set_del(struct ip_set *set, struct set_elem *e)
{ {
set->elements--;
list_del_rcu(&e->list); list_del_rcu(&e->list);
call_rcu(&e->rcu, __list_set_del_rcu); call_rcu(&e->rcu, __list_set_del_rcu);
} }
...@@ -309,6 +310,7 @@ list_set_uadd(struct ip_set *set, void *value, const struct ip_set_ext *ext, ...@@ -309,6 +310,7 @@ list_set_uadd(struct ip_set *set, void *value, const struct ip_set_ext *ext,
list_add_rcu(&e->list, &prev->list); list_add_rcu(&e->list, &prev->list);
else else
list_add_tail_rcu(&e->list, &map->members); list_add_tail_rcu(&e->list, &map->members);
set->elements++;
return 0; return 0;
} }
...@@ -419,6 +421,7 @@ list_set_flush(struct ip_set *set) ...@@ -419,6 +421,7 @@ list_set_flush(struct ip_set *set)
list_for_each_entry_safe(e, n, &map->members, list) list_for_each_entry_safe(e, n, &map->members, list)
list_set_del(set, e); list_set_del(set, e);
set->elements = 0;
} }
static void static void
...@@ -471,7 +474,8 @@ list_set_head(struct ip_set *set, struct sk_buff *skb) ...@@ -471,7 +474,8 @@ list_set_head(struct ip_set *set, struct sk_buff *skb)
goto nla_put_failure; goto nla_put_failure;
if (nla_put_net32(skb, IPSET_ATTR_SIZE, htonl(map->size)) || if (nla_put_net32(skb, IPSET_ATTR_SIZE, htonl(map->size)) ||
nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) || nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) ||
nla_put_net32(skb, IPSET_ATTR_MEMSIZE, htonl(memsize))) nla_put_net32(skb, IPSET_ATTR_MEMSIZE, htonl(memsize)) ||
nla_put_net32(skb, IPSET_ATTR_ELEMENTS, htonl(set->elements)))
goto nla_put_failure; goto nla_put_failure;
if (unlikely(ip_set_put_flags(skb, set))) if (unlikely(ip_set_put_flags(skb, set)))
goto nla_put_failure; goto nla_put_failure;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment