Commit b637e498 authored by Pravin B Shelar's avatar Pravin B Shelar Committed by Jesse Gross

openvswitch: Move mega-flow list out of rehashing struct.

ovs-flow rehash does not touch mega flow list. Following patch
moves it dp struct datapath.  Avoid one extra indirection for
accessing mega-flow list head on every packet receive.
Signed-off-by: default avatarPravin B Shelar <pshelar@nicira.com>
Signed-off-by: default avatarJesse Gross <jesse@nicira.com>
parent e6445719
...@@ -59,8 +59,6 @@ ...@@ -59,8 +59,6 @@
#include "vport-internal_dev.h" #include "vport-internal_dev.h"
#include "vport-netdev.h" #include "vport-netdev.h"
#define REHASH_FLOW_INTERVAL (10 * 60 * HZ)
int ovs_net_id __read_mostly; int ovs_net_id __read_mostly;
static void ovs_notify(struct sk_buff *skb, struct genl_info *info, static void ovs_notify(struct sk_buff *skb, struct genl_info *info,
...@@ -163,7 +161,7 @@ static void destroy_dp_rcu(struct rcu_head *rcu) ...@@ -163,7 +161,7 @@ static void destroy_dp_rcu(struct rcu_head *rcu)
{ {
struct datapath *dp = container_of(rcu, struct datapath, rcu); struct datapath *dp = container_of(rcu, struct datapath, rcu);
ovs_flow_tbl_destroy((__force struct flow_table *)dp->table, false); ovs_flow_tbl_destroy(&dp->table, false);
free_percpu(dp->stats_percpu); free_percpu(dp->stats_percpu);
release_net(ovs_dp_get_net(dp)); release_net(ovs_dp_get_net(dp));
kfree(dp->ports); kfree(dp->ports);
...@@ -235,7 +233,7 @@ void ovs_dp_process_received_packet(struct vport *p, struct sk_buff *skb) ...@@ -235,7 +233,7 @@ void ovs_dp_process_received_packet(struct vport *p, struct sk_buff *skb)
} }
/* Look up flow. */ /* Look up flow. */
flow = ovs_flow_tbl_lookup(rcu_dereference(dp->table), &key); flow = ovs_flow_tbl_lookup(&dp->table, &key);
if (unlikely(!flow)) { if (unlikely(!flow)) {
struct dp_upcall_info upcall; struct dp_upcall_info upcall;
...@@ -453,23 +451,6 @@ static int queue_userspace_packet(struct net *net, int dp_ifindex, ...@@ -453,23 +451,6 @@ static int queue_userspace_packet(struct net *net, int dp_ifindex,
return err; return err;
} }
/* Called with ovs_mutex. */
static int flush_flows(struct datapath *dp)
{
struct flow_table *old_table;
struct flow_table *new_table;
old_table = ovsl_dereference(dp->table);
new_table = ovs_flow_tbl_alloc(TBL_MIN_BUCKETS);
if (!new_table)
return -ENOMEM;
rcu_assign_pointer(dp->table, new_table);
ovs_flow_tbl_destroy(old_table, true);
return 0;
}
static void clear_stats(struct sw_flow *flow) static void clear_stats(struct sw_flow *flow)
{ {
flow->used = 0; flow->used = 0;
...@@ -584,11 +565,9 @@ static struct genl_ops dp_packet_genl_ops[] = { ...@@ -584,11 +565,9 @@ static struct genl_ops dp_packet_genl_ops[] = {
static void get_dp_stats(struct datapath *dp, struct ovs_dp_stats *stats) static void get_dp_stats(struct datapath *dp, struct ovs_dp_stats *stats)
{ {
struct flow_table *table;
int i; int i;
table = rcu_dereference_check(dp->table, lockdep_ovsl_is_held()); stats->n_flows = ovs_flow_tbl_count(&dp->table);
stats->n_flows = ovs_flow_tbl_count(table);
stats->n_hit = stats->n_missed = stats->n_lost = 0; stats->n_hit = stats->n_missed = stats->n_lost = 0;
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
...@@ -773,7 +752,6 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info) ...@@ -773,7 +752,6 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
struct sw_flow_mask mask; struct sw_flow_mask mask;
struct sk_buff *reply; struct sk_buff *reply;
struct datapath *dp; struct datapath *dp;
struct flow_table *table;
struct sw_flow_actions *acts = NULL; struct sw_flow_actions *acts = NULL;
struct sw_flow_match match; struct sw_flow_match match;
int error; int error;
...@@ -814,12 +792,9 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info) ...@@ -814,12 +792,9 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
if (!dp) if (!dp)
goto err_unlock_ovs; goto err_unlock_ovs;
table = ovsl_dereference(dp->table);
/* Check if this is a duplicate flow */ /* Check if this is a duplicate flow */
flow = ovs_flow_tbl_lookup(table, &key); flow = ovs_flow_tbl_lookup(&dp->table, &key);
if (!flow) { if (!flow) {
struct flow_table *new_table = NULL;
struct sw_flow_mask *mask_p; struct sw_flow_mask *mask_p;
/* Bail out if we're not allowed to create a new flow. */ /* Bail out if we're not allowed to create a new flow. */
...@@ -827,19 +802,6 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info) ...@@ -827,19 +802,6 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
if (info->genlhdr->cmd == OVS_FLOW_CMD_SET) if (info->genlhdr->cmd == OVS_FLOW_CMD_SET)
goto err_unlock_ovs; goto err_unlock_ovs;
/* Expand table, if necessary, to make room. */
if (ovs_flow_tbl_need_to_expand(table))
new_table = ovs_flow_tbl_expand(table);
else if (time_after(jiffies, dp->last_rehash + REHASH_FLOW_INTERVAL))
new_table = ovs_flow_tbl_rehash(table);
if (new_table && !IS_ERR(new_table)) {
rcu_assign_pointer(dp->table, new_table);
ovs_flow_tbl_destroy(table, true);
table = ovsl_dereference(dp->table);
dp->last_rehash = jiffies;
}
/* Allocate flow. */ /* Allocate flow. */
flow = ovs_flow_alloc(); flow = ovs_flow_alloc();
if (IS_ERR(flow)) { if (IS_ERR(flow)) {
...@@ -852,7 +814,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info) ...@@ -852,7 +814,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
flow->unmasked_key = key; flow->unmasked_key = key;
/* Make sure mask is unique in the system */ /* Make sure mask is unique in the system */
mask_p = ovs_sw_flow_mask_find(table, &mask); mask_p = ovs_sw_flow_mask_find(&dp->table, &mask);
if (!mask_p) { if (!mask_p) {
/* Allocate a new mask if none exsits. */ /* Allocate a new mask if none exsits. */
mask_p = ovs_sw_flow_mask_alloc(); mask_p = ovs_sw_flow_mask_alloc();
...@@ -860,7 +822,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info) ...@@ -860,7 +822,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
goto err_flow_free; goto err_flow_free;
mask_p->key = mask.key; mask_p->key = mask.key;
mask_p->range = mask.range; mask_p->range = mask.range;
ovs_sw_flow_mask_insert(table, mask_p); ovs_sw_flow_mask_insert(&dp->table, mask_p);
} }
ovs_sw_flow_mask_add_ref(mask_p); ovs_sw_flow_mask_add_ref(mask_p);
...@@ -868,7 +830,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info) ...@@ -868,7 +830,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
rcu_assign_pointer(flow->sf_acts, acts); rcu_assign_pointer(flow->sf_acts, acts);
/* Put flow in bucket. */ /* Put flow in bucket. */
ovs_flow_tbl_insert(table, flow); ovs_flow_tbl_insert(&dp->table, flow);
reply = ovs_flow_cmd_build_info(flow, dp, info->snd_portid, reply = ovs_flow_cmd_build_info(flow, dp, info->snd_portid,
info->snd_seq, OVS_FLOW_CMD_NEW); info->snd_seq, OVS_FLOW_CMD_NEW);
...@@ -936,7 +898,6 @@ static int ovs_flow_cmd_get(struct sk_buff *skb, struct genl_info *info) ...@@ -936,7 +898,6 @@ static int ovs_flow_cmd_get(struct sk_buff *skb, struct genl_info *info)
struct sk_buff *reply; struct sk_buff *reply;
struct sw_flow *flow; struct sw_flow *flow;
struct datapath *dp; struct datapath *dp;
struct flow_table *table;
struct sw_flow_match match; struct sw_flow_match match;
int err; int err;
...@@ -957,8 +918,7 @@ static int ovs_flow_cmd_get(struct sk_buff *skb, struct genl_info *info) ...@@ -957,8 +918,7 @@ static int ovs_flow_cmd_get(struct sk_buff *skb, struct genl_info *info)
goto unlock; goto unlock;
} }
table = ovsl_dereference(dp->table); flow = ovs_flow_tbl_lookup(&dp->table, &key);
flow = ovs_flow_tbl_lookup(table, &key);
if (!flow || !ovs_flow_cmp_unmasked_key(flow, &match)) { if (!flow || !ovs_flow_cmp_unmasked_key(flow, &match)) {
err = -ENOENT; err = -ENOENT;
goto unlock; goto unlock;
...@@ -986,7 +946,6 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info) ...@@ -986,7 +946,6 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info)
struct sk_buff *reply; struct sk_buff *reply;
struct sw_flow *flow; struct sw_flow *flow;
struct datapath *dp; struct datapath *dp;
struct flow_table *table;
struct sw_flow_match match; struct sw_flow_match match;
int err; int err;
...@@ -998,7 +957,7 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info) ...@@ -998,7 +957,7 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info)
} }
if (!a[OVS_FLOW_ATTR_KEY]) { if (!a[OVS_FLOW_ATTR_KEY]) {
err = flush_flows(dp); err = ovs_flow_tbl_flush(&dp->table);
goto unlock; goto unlock;
} }
...@@ -1007,8 +966,7 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info) ...@@ -1007,8 +966,7 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info)
if (err) if (err)
goto unlock; goto unlock;
table = ovsl_dereference(dp->table); flow = ovs_flow_tbl_lookup(&dp->table, &key);
flow = ovs_flow_tbl_lookup(table, &key);
if (!flow || !ovs_flow_cmp_unmasked_key(flow, &match)) { if (!flow || !ovs_flow_cmp_unmasked_key(flow, &match)) {
err = -ENOENT; err = -ENOENT;
goto unlock; goto unlock;
...@@ -1020,7 +978,7 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info) ...@@ -1020,7 +978,7 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info)
goto unlock; goto unlock;
} }
ovs_flow_tbl_remove(table, flow); ovs_flow_tbl_remove(&dp->table, flow);
err = ovs_flow_cmd_fill_info(flow, dp, reply, info->snd_portid, err = ovs_flow_cmd_fill_info(flow, dp, reply, info->snd_portid,
info->snd_seq, 0, OVS_FLOW_CMD_DEL); info->snd_seq, 0, OVS_FLOW_CMD_DEL);
...@@ -1039,8 +997,8 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info) ...@@ -1039,8 +997,8 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info)
static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
{ {
struct ovs_header *ovs_header = genlmsg_data(nlmsg_data(cb->nlh)); struct ovs_header *ovs_header = genlmsg_data(nlmsg_data(cb->nlh));
struct table_instance *ti;
struct datapath *dp; struct datapath *dp;
struct flow_table *table;
rcu_read_lock(); rcu_read_lock();
dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex); dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
...@@ -1049,14 +1007,14 @@ static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) ...@@ -1049,14 +1007,14 @@ static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
return -ENODEV; return -ENODEV;
} }
table = rcu_dereference(dp->table); ti = rcu_dereference(dp->table.ti);
for (;;) { for (;;) {
struct sw_flow *flow; struct sw_flow *flow;
u32 bucket, obj; u32 bucket, obj;
bucket = cb->args[0]; bucket = cb->args[0];
obj = cb->args[1]; obj = cb->args[1];
flow = ovs_flow_tbl_dump_next(table, &bucket, &obj); flow = ovs_flow_tbl_dump_next(ti, &bucket, &obj);
if (!flow) if (!flow)
break; break;
...@@ -1220,9 +1178,8 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info) ...@@ -1220,9 +1178,8 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
ovs_dp_set_net(dp, hold_net(sock_net(skb->sk))); ovs_dp_set_net(dp, hold_net(sock_net(skb->sk)));
/* Allocate table. */ /* Allocate table. */
err = -ENOMEM; err = ovs_flow_tbl_init(&dp->table);
rcu_assign_pointer(dp->table, ovs_flow_tbl_alloc(TBL_MIN_BUCKETS)); if (err)
if (!dp->table)
goto err_free_dp; goto err_free_dp;
dp->stats_percpu = alloc_percpu(struct dp_stats_percpu); dp->stats_percpu = alloc_percpu(struct dp_stats_percpu);
...@@ -1279,7 +1236,7 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info) ...@@ -1279,7 +1236,7 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
err_destroy_percpu: err_destroy_percpu:
free_percpu(dp->stats_percpu); free_percpu(dp->stats_percpu);
err_destroy_table: err_destroy_table:
ovs_flow_tbl_destroy(ovsl_dereference(dp->table), false); ovs_flow_tbl_destroy(&dp->table, false);
err_free_dp: err_free_dp:
release_net(ovs_dp_get_net(dp)); release_net(ovs_dp_get_net(dp));
kfree(dp); kfree(dp);
......
...@@ -58,12 +58,11 @@ struct dp_stats_percpu { ...@@ -58,12 +58,11 @@ struct dp_stats_percpu {
* struct datapath - datapath for flow-based packet switching * struct datapath - datapath for flow-based packet switching
* @rcu: RCU callback head for deferred destruction. * @rcu: RCU callback head for deferred destruction.
* @list_node: Element in global 'dps' list. * @list_node: Element in global 'dps' list.
* @table: Current flow table. Protected by ovs_mutex and RCU. * @table: flow table.
* @ports: Hash table for ports. %OVSP_LOCAL port always exists. Protected by * @ports: Hash table for ports. %OVSP_LOCAL port always exists. Protected by
* ovs_mutex and RCU. * ovs_mutex and RCU.
* @stats_percpu: Per-CPU datapath statistics. * @stats_percpu: Per-CPU datapath statistics.
* @net: Reference to net namespace. * @net: Reference to net namespace.
* @last_rehash: Timestamp of last rehash.
* *
* Context: See the comment on locking at the top of datapath.c for additional * Context: See the comment on locking at the top of datapath.c for additional
* locking information. * locking information.
...@@ -73,7 +72,7 @@ struct datapath { ...@@ -73,7 +72,7 @@ struct datapath {
struct list_head list_node; struct list_head list_node;
/* Flow table. */ /* Flow table. */
struct flow_table __rcu *table; struct flow_table table;
/* Switch ports. */ /* Switch ports. */
struct hlist_head *ports; struct hlist_head *ports;
...@@ -85,7 +84,6 @@ struct datapath { ...@@ -85,7 +84,6 @@ struct datapath {
/* Network namespace ref. */ /* Network namespace ref. */
struct net *net; struct net *net;
#endif #endif
unsigned long last_rehash;
}; };
/** /**
......
...@@ -44,6 +44,11 @@ ...@@ -44,6 +44,11 @@
#include <net/ipv6.h> #include <net/ipv6.h>
#include <net/ndisc.h> #include <net/ndisc.h>
#include "datapath.h"
#define TBL_MIN_BUCKETS 1024
#define REHASH_INTERVAL (10 * 60 * HZ)
static struct kmem_cache *flow_cache; static struct kmem_cache *flow_cache;
static u16 range_n_bytes(const struct sw_flow_key_range *range) static u16 range_n_bytes(const struct sw_flow_key_range *range)
...@@ -82,6 +87,11 @@ struct sw_flow *ovs_flow_alloc(void) ...@@ -82,6 +87,11 @@ struct sw_flow *ovs_flow_alloc(void)
return flow; return flow;
} }
int ovs_flow_tbl_count(struct flow_table *table)
{
return table->count;
}
static struct flex_array *alloc_buckets(unsigned int n_buckets) static struct flex_array *alloc_buckets(unsigned int n_buckets)
{ {
struct flex_array *buckets; struct flex_array *buckets;
...@@ -136,18 +146,18 @@ static void free_buckets(struct flex_array *buckets) ...@@ -136,18 +146,18 @@ static void free_buckets(struct flex_array *buckets)
flex_array_free(buckets); flex_array_free(buckets);
} }
static void __flow_tbl_destroy(struct flow_table *table) static void __table_instance_destroy(struct table_instance *ti)
{ {
int i; int i;
if (table->keep_flows) if (ti->keep_flows)
goto skip_flows; goto skip_flows;
for (i = 0; i < table->n_buckets; i++) { for (i = 0; i < ti->n_buckets; i++) {
struct sw_flow *flow; struct sw_flow *flow;
struct hlist_head *head = flex_array_get(table->buckets, i); struct hlist_head *head = flex_array_get(ti->buckets, i);
struct hlist_node *n; struct hlist_node *n;
int ver = table->node_ver; int ver = ti->node_ver;
hlist_for_each_entry_safe(flow, n, head, hash_node[ver]) { hlist_for_each_entry_safe(flow, n, head, hash_node[ver]) {
hlist_del(&flow->hash_node[ver]); hlist_del(&flow->hash_node[ver]);
...@@ -155,74 +165,74 @@ static void __flow_tbl_destroy(struct flow_table *table) ...@@ -155,74 +165,74 @@ static void __flow_tbl_destroy(struct flow_table *table)
} }
} }
BUG_ON(!list_empty(table->mask_list));
kfree(table->mask_list);
skip_flows: skip_flows:
free_buckets(table->buckets); free_buckets(ti->buckets);
kfree(table); kfree(ti);
} }
static struct flow_table *__flow_tbl_alloc(int new_size) static struct table_instance *table_instance_alloc(int new_size)
{ {
struct flow_table *table = kmalloc(sizeof(*table), GFP_KERNEL); struct table_instance *ti = kmalloc(sizeof(*ti), GFP_KERNEL);
if (!table) if (!ti)
return NULL; return NULL;
table->buckets = alloc_buckets(new_size); ti->buckets = alloc_buckets(new_size);
if (!table->buckets) { if (!ti->buckets) {
kfree(table); kfree(ti);
return NULL; return NULL;
} }
table->n_buckets = new_size; ti->n_buckets = new_size;
table->count = 0; ti->node_ver = 0;
table->node_ver = 0; ti->keep_flows = false;
table->keep_flows = false; get_random_bytes(&ti->hash_seed, sizeof(u32));
get_random_bytes(&table->hash_seed, sizeof(u32));
table->mask_list = NULL;
return table; return ti;
} }
struct flow_table *ovs_flow_tbl_alloc(int new_size) int ovs_flow_tbl_init(struct flow_table *table)
{ {
struct flow_table *table = __flow_tbl_alloc(new_size); struct table_instance *ti;
if (!table) ti = table_instance_alloc(TBL_MIN_BUCKETS);
return NULL;
table->mask_list = kmalloc(sizeof(struct list_head), GFP_KERNEL); if (!ti)
if (!table->mask_list) { return -ENOMEM;
table->keep_flows = true;
__flow_tbl_destroy(table);
return NULL;
}
INIT_LIST_HEAD(table->mask_list);
return table; rcu_assign_pointer(table->ti, ti);
INIT_LIST_HEAD(&table->mask_list);
table->last_rehash = jiffies;
table->count = 0;
return 0;
} }
static void flow_tbl_destroy_rcu_cb(struct rcu_head *rcu) static void flow_tbl_destroy_rcu_cb(struct rcu_head *rcu)
{ {
struct flow_table *table = container_of(rcu, struct flow_table, rcu); struct table_instance *ti = container_of(rcu, struct table_instance, rcu);
__flow_tbl_destroy(table); __table_instance_destroy(ti);
} }
void ovs_flow_tbl_destroy(struct flow_table *table, bool deferred) static void table_instance_destroy(struct table_instance *ti, bool deferred)
{ {
if (!table) if (!ti)
return; return;
if (deferred) if (deferred)
call_rcu(&table->rcu, flow_tbl_destroy_rcu_cb); call_rcu(&ti->rcu, flow_tbl_destroy_rcu_cb);
else else
__flow_tbl_destroy(table); __table_instance_destroy(ti);
}
void ovs_flow_tbl_destroy(struct flow_table *table, bool deferred)
{
struct table_instance *ti = ovsl_dereference(table->ti);
table_instance_destroy(ti, deferred);
} }
struct sw_flow *ovs_flow_tbl_dump_next(struct flow_table *table, struct sw_flow *ovs_flow_tbl_dump_next(struct table_instance *ti,
u32 *bucket, u32 *last) u32 *bucket, u32 *last)
{ {
struct sw_flow *flow; struct sw_flow *flow;
...@@ -230,10 +240,10 @@ struct sw_flow *ovs_flow_tbl_dump_next(struct flow_table *table, ...@@ -230,10 +240,10 @@ struct sw_flow *ovs_flow_tbl_dump_next(struct flow_table *table,
int ver; int ver;
int i; int i;
ver = table->node_ver; ver = ti->node_ver;
while (*bucket < table->n_buckets) { while (*bucket < ti->n_buckets) {
i = 0; i = 0;
head = flex_array_get(table->buckets, *bucket); head = flex_array_get(ti->buckets, *bucket);
hlist_for_each_entry_rcu(flow, head, hash_node[ver]) { hlist_for_each_entry_rcu(flow, head, hash_node[ver]) {
if (i < *last) { if (i < *last) {
i++; i++;
...@@ -249,25 +259,23 @@ struct sw_flow *ovs_flow_tbl_dump_next(struct flow_table *table, ...@@ -249,25 +259,23 @@ struct sw_flow *ovs_flow_tbl_dump_next(struct flow_table *table,
return NULL; return NULL;
} }
static struct hlist_head *find_bucket(struct flow_table *table, u32 hash) static struct hlist_head *find_bucket(struct table_instance *ti, u32 hash)
{ {
hash = jhash_1word(hash, table->hash_seed); hash = jhash_1word(hash, ti->hash_seed);
return flex_array_get(table->buckets, return flex_array_get(ti->buckets,
(hash & (table->n_buckets - 1))); (hash & (ti->n_buckets - 1)));
} }
static void __tbl_insert(struct flow_table *table, struct sw_flow *flow) static void table_instance_insert(struct table_instance *ti, struct sw_flow *flow)
{ {
struct hlist_head *head; struct hlist_head *head;
head = find_bucket(table, flow->hash); head = find_bucket(ti, flow->hash);
hlist_add_head_rcu(&flow->hash_node[table->node_ver], head); hlist_add_head_rcu(&flow->hash_node[ti->node_ver], head);
table->count++;
} }
static void flow_table_copy_flows(struct flow_table *old, static void flow_table_copy_flows(struct table_instance *old,
struct flow_table *new) struct table_instance *new)
{ {
int old_ver; int old_ver;
int i; int i;
...@@ -283,35 +291,42 @@ static void flow_table_copy_flows(struct flow_table *old, ...@@ -283,35 +291,42 @@ static void flow_table_copy_flows(struct flow_table *old,
head = flex_array_get(old->buckets, i); head = flex_array_get(old->buckets, i);
hlist_for_each_entry(flow, head, hash_node[old_ver]) hlist_for_each_entry(flow, head, hash_node[old_ver])
__tbl_insert(new, flow); table_instance_insert(new, flow);
} }
new->mask_list = old->mask_list;
old->keep_flows = true; old->keep_flows = true;
} }
static struct flow_table *__flow_tbl_rehash(struct flow_table *table, static struct table_instance *table_instance_rehash(struct table_instance *ti,
int n_buckets) int n_buckets)
{ {
struct flow_table *new_table; struct table_instance *new_ti;
new_table = __flow_tbl_alloc(n_buckets); new_ti = table_instance_alloc(n_buckets);
if (!new_table) if (!new_ti)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
flow_table_copy_flows(table, new_table); flow_table_copy_flows(ti, new_ti);
return new_table; return new_ti;
} }
struct flow_table *ovs_flow_tbl_rehash(struct flow_table *table) int ovs_flow_tbl_flush(struct flow_table *flow_table)
{ {
return __flow_tbl_rehash(table, table->n_buckets); struct table_instance *old_ti;
} struct table_instance *new_ti;
struct flow_table *ovs_flow_tbl_expand(struct flow_table *table) old_ti = ovsl_dereference(flow_table->ti);
{ new_ti = table_instance_alloc(TBL_MIN_BUCKETS);
return __flow_tbl_rehash(table, table->n_buckets * 2); if (!new_ti)
return -ENOMEM;
rcu_assign_pointer(flow_table->ti, new_ti);
flow_table->last_rehash = jiffies;
flow_table->count = 0;
table_instance_destroy(old_ti, true);
return 0;
} }
static u32 flow_hash(const struct sw_flow_key *key, int key_start, static u32 flow_hash(const struct sw_flow_key *key, int key_start,
...@@ -367,7 +382,7 @@ bool ovs_flow_cmp_unmasked_key(const struct sw_flow *flow, ...@@ -367,7 +382,7 @@ bool ovs_flow_cmp_unmasked_key(const struct sw_flow *flow,
return cmp_key(&flow->unmasked_key, key, key_start, key_end); return cmp_key(&flow->unmasked_key, key, key_start, key_end);
} }
static struct sw_flow *masked_flow_lookup(struct flow_table *table, static struct sw_flow *masked_flow_lookup(struct table_instance *ti,
const struct sw_flow_key *unmasked, const struct sw_flow_key *unmasked,
struct sw_flow_mask *mask) struct sw_flow_mask *mask)
{ {
...@@ -380,8 +395,8 @@ static struct sw_flow *masked_flow_lookup(struct flow_table *table, ...@@ -380,8 +395,8 @@ static struct sw_flow *masked_flow_lookup(struct flow_table *table,
ovs_flow_mask_key(&masked_key, unmasked, mask); ovs_flow_mask_key(&masked_key, unmasked, mask);
hash = flow_hash(&masked_key, key_start, key_end); hash = flow_hash(&masked_key, key_start, key_end);
head = find_bucket(table, hash); head = find_bucket(ti, hash);
hlist_for_each_entry_rcu(flow, head, hash_node[table->node_ver]) { hlist_for_each_entry_rcu(flow, head, hash_node[ti->node_ver]) {
if (flow->mask == mask && if (flow->mask == mask &&
flow_cmp_masked_key(flow, &masked_key, flow_cmp_masked_key(flow, &masked_key,
key_start, key_end)) key_start, key_end))
...@@ -393,29 +408,55 @@ static struct sw_flow *masked_flow_lookup(struct flow_table *table, ...@@ -393,29 +408,55 @@ static struct sw_flow *masked_flow_lookup(struct flow_table *table,
struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *tbl, struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *tbl,
const struct sw_flow_key *key) const struct sw_flow_key *key)
{ {
struct sw_flow *flow = NULL; struct table_instance *ti = rcu_dereference(tbl->ti);
struct sw_flow_mask *mask; struct sw_flow_mask *mask;
struct sw_flow *flow;
list_for_each_entry_rcu(mask, tbl->mask_list, list) { list_for_each_entry_rcu(mask, &tbl->mask_list, list) {
flow = masked_flow_lookup(tbl, key, mask); flow = masked_flow_lookup(ti, key, mask);
if (flow) /* Found */ if (flow) /* Found */
break; return flow;
} }
return NULL;
}
return flow; static struct table_instance *table_instance_expand(struct table_instance *ti)
{
return table_instance_rehash(ti, ti->n_buckets * 2);
} }
void ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow) void ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow)
{ {
struct table_instance *ti = NULL;
struct table_instance *new_ti = NULL;
ti = ovsl_dereference(table->ti);
/* Expand table, if necessary, to make room. */
if (table->count > ti->n_buckets)
new_ti = table_instance_expand(ti);
else if (time_after(jiffies, table->last_rehash + REHASH_INTERVAL))
new_ti = table_instance_rehash(ti, ti->n_buckets);
if (new_ti && !IS_ERR(new_ti)) {
rcu_assign_pointer(table->ti, new_ti);
ovs_flow_tbl_destroy(table, true);
ti = ovsl_dereference(table->ti);
table->last_rehash = jiffies;
}
flow->hash = flow_hash(&flow->key, flow->mask->range.start, flow->hash = flow_hash(&flow->key, flow->mask->range.start,
flow->mask->range.end); flow->mask->range.end);
__tbl_insert(table, flow); table_instance_insert(ti, flow);
table->count++;
} }
void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow) void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow)
{ {
struct table_instance *ti = ovsl_dereference(table->ti);
BUG_ON(table->count == 0); BUG_ON(table->count == 0);
hlist_del_rcu(&flow->hash_node[table->node_ver]); hlist_del_rcu(&flow->hash_node[ti->node_ver]);
table->count--; table->count--;
} }
...@@ -475,7 +516,7 @@ struct sw_flow_mask *ovs_sw_flow_mask_find(const struct flow_table *tbl, ...@@ -475,7 +516,7 @@ struct sw_flow_mask *ovs_sw_flow_mask_find(const struct flow_table *tbl,
{ {
struct list_head *ml; struct list_head *ml;
list_for_each(ml, tbl->mask_list) { list_for_each(ml, &tbl->mask_list) {
struct sw_flow_mask *m; struct sw_flow_mask *m;
m = container_of(ml, struct sw_flow_mask, list); m = container_of(ml, struct sw_flow_mask, list);
if (mask_equal(mask, m)) if (mask_equal(mask, m))
...@@ -492,7 +533,7 @@ struct sw_flow_mask *ovs_sw_flow_mask_find(const struct flow_table *tbl, ...@@ -492,7 +533,7 @@ struct sw_flow_mask *ovs_sw_flow_mask_find(const struct flow_table *tbl,
*/ */
void ovs_sw_flow_mask_insert(struct flow_table *tbl, struct sw_flow_mask *mask) void ovs_sw_flow_mask_insert(struct flow_table *tbl, struct sw_flow_mask *mask)
{ {
list_add_rcu(&mask->list, tbl->mask_list); list_add_rcu(&mask->list, &tbl->mask_list);
} }
/* Initializes the flow module. /* Initializes the flow module.
......
...@@ -36,42 +36,36 @@ ...@@ -36,42 +36,36 @@
#include "flow.h" #include "flow.h"
#define TBL_MIN_BUCKETS 1024 struct table_instance {
struct flow_table {
struct flex_array *buckets; struct flex_array *buckets;
unsigned int count, n_buckets; unsigned int n_buckets;
struct rcu_head rcu; struct rcu_head rcu;
struct list_head *mask_list;
int node_ver; int node_ver;
u32 hash_seed; u32 hash_seed;
bool keep_flows; bool keep_flows;
}; };
struct flow_table {
struct table_instance __rcu *ti;
struct list_head mask_list;
unsigned long last_rehash;
unsigned int count;
};
int ovs_flow_init(void); int ovs_flow_init(void);
void ovs_flow_exit(void); void ovs_flow_exit(void);
struct sw_flow *ovs_flow_alloc(void); struct sw_flow *ovs_flow_alloc(void);
void ovs_flow_free(struct sw_flow *, bool deferred); void ovs_flow_free(struct sw_flow *, bool deferred);
static inline int ovs_flow_tbl_count(struct flow_table *table) int ovs_flow_tbl_init(struct flow_table *);
{ int ovs_flow_tbl_count(struct flow_table *table);
return table->count;
}
static inline int ovs_flow_tbl_need_to_expand(struct flow_table *table)
{
return (table->count > table->n_buckets);
}
struct flow_table *ovs_flow_tbl_alloc(int new_size);
struct flow_table *ovs_flow_tbl_expand(struct flow_table *table);
struct flow_table *ovs_flow_tbl_rehash(struct flow_table *table);
void ovs_flow_tbl_destroy(struct flow_table *table, bool deferred); void ovs_flow_tbl_destroy(struct flow_table *table, bool deferred);
int ovs_flow_tbl_flush(struct flow_table *flow_table);
void ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow); void ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow);
void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow); void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow);
struct sw_flow *ovs_flow_tbl_dump_next(struct flow_table *table, struct sw_flow *ovs_flow_tbl_dump_next(struct table_instance *table,
u32 *bucket, u32 *idx); u32 *bucket, u32 *idx);
struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *, struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *,
const struct sw_flow_key *); const struct sw_flow_key *);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment