Commit 96442e42 authored by Jason Wang's avatar Jason Wang Committed by David S. Miller

tuntap: choose the txq based on rxq

This patch implements a simple multiqueue flow steering policy - tx follows rx
for tun/tap. The idea is simple, it just choose the txq based on which rxq it
comes. The flow were identified through the rxhash of a skb, and the hash to
queue mapping were recorded in a hlist with an ageing timer to retire the
mapping. The mapping were created when tun receives packet from userspace, and
was quired in .ndo_select_queue().

I run co-current TCP_CRR test and didn't see any mapping manipulation helpers in
perf top, so the overhead could be negelected.
Signed-off-by: default avatarJason Wang <jasowang@redhat.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent cde8b15f
...@@ -115,6 +115,8 @@ struct tap_filter { ...@@ -115,6 +115,8 @@ struct tap_filter {
*/ */
#define MAX_TAP_QUEUES 1024 #define MAX_TAP_QUEUES 1024
#define TUN_FLOW_EXPIRE (3 * HZ)
/* A tun_file connects an open character device to a tuntap netdevice. It /* A tun_file connects an open character device to a tuntap netdevice. It
* also contains all socket related strctures (except sock_fprog and tap_filter) * also contains all socket related strctures (except sock_fprog and tap_filter)
* to serve as one transmit queue for tuntap device. The sock_fprog and * to serve as one transmit queue for tuntap device. The sock_fprog and
...@@ -138,6 +140,18 @@ struct tun_file { ...@@ -138,6 +140,18 @@ struct tun_file {
u16 queue_index; u16 queue_index;
}; };
struct tun_flow_entry {
struct hlist_node hash_link;
struct rcu_head rcu;
struct tun_struct *tun;
u32 rxhash;
int queue_index;
unsigned long updated;
};
#define TUN_NUM_FLOW_ENTRIES 1024
/* Since the socket were moved to tun_file, to preserve the behavior of persist /* Since the socket were moved to tun_file, to preserve the behavior of persist
* device, socket fileter, sndbuf and vnet header size were restore when the * device, socket fileter, sndbuf and vnet header size were restore when the
* file were attached to a persist device. * file were attached to a persist device.
...@@ -163,8 +177,164 @@ struct tun_struct { ...@@ -163,8 +177,164 @@ struct tun_struct {
#ifdef TUN_DEBUG #ifdef TUN_DEBUG
int debug; int debug;
#endif #endif
spinlock_t lock;
struct kmem_cache *flow_cache;
struct hlist_head flows[TUN_NUM_FLOW_ENTRIES];
struct timer_list flow_gc_timer;
unsigned long ageing_time;
}; };
static inline u32 tun_hashfn(u32 rxhash)
{
return rxhash & 0x3ff;
}
static struct tun_flow_entry *tun_flow_find(struct hlist_head *head, u32 rxhash)
{
struct tun_flow_entry *e;
struct hlist_node *n;
hlist_for_each_entry_rcu(e, n, head, hash_link) {
if (e->rxhash == rxhash)
return e;
}
return NULL;
}
static struct tun_flow_entry *tun_flow_create(struct tun_struct *tun,
struct hlist_head *head,
u32 rxhash, u16 queue_index)
{
struct tun_flow_entry *e = kmem_cache_alloc(tun->flow_cache,
GFP_ATOMIC);
if (e) {
tun_debug(KERN_INFO, tun, "create flow: hash %u index %u\n",
rxhash, queue_index);
e->updated = jiffies;
e->rxhash = rxhash;
e->queue_index = queue_index;
e->tun = tun;
hlist_add_head_rcu(&e->hash_link, head);
}
return e;
}
static void tun_flow_free(struct rcu_head *head)
{
struct tun_flow_entry *e
= container_of(head, struct tun_flow_entry, rcu);
kmem_cache_free(e->tun->flow_cache, e);
}
static void tun_flow_delete(struct tun_struct *tun, struct tun_flow_entry *e)
{
tun_debug(KERN_INFO, tun, "delete flow: hash %u index %u\n",
e->rxhash, e->queue_index);
hlist_del_rcu(&e->hash_link);
call_rcu(&e->rcu, tun_flow_free);
}
static void tun_flow_flush(struct tun_struct *tun)
{
int i;
spin_lock_bh(&tun->lock);
for (i = 0; i < TUN_NUM_FLOW_ENTRIES; i++) {
struct tun_flow_entry *e;
struct hlist_node *h, *n;
hlist_for_each_entry_safe(e, h, n, &tun->flows[i], hash_link)
tun_flow_delete(tun, e);
}
spin_unlock_bh(&tun->lock);
}
static void tun_flow_delete_by_queue(struct tun_struct *tun, u16 queue_index)
{
int i;
spin_lock_bh(&tun->lock);
for (i = 0; i < TUN_NUM_FLOW_ENTRIES; i++) {
struct tun_flow_entry *e;
struct hlist_node *h, *n;
hlist_for_each_entry_safe(e, h, n, &tun->flows[i], hash_link) {
if (e->queue_index == queue_index)
tun_flow_delete(tun, e);
}
}
spin_unlock_bh(&tun->lock);
}
static void tun_flow_cleanup(unsigned long data)
{
struct tun_struct *tun = (struct tun_struct *)data;
unsigned long delay = tun->ageing_time;
unsigned long next_timer = jiffies + delay;
unsigned long count = 0;
int i;
tun_debug(KERN_INFO, tun, "tun_flow_cleanup\n");
spin_lock_bh(&tun->lock);
for (i = 0; i < TUN_NUM_FLOW_ENTRIES; i++) {
struct tun_flow_entry *e;
struct hlist_node *h, *n;
hlist_for_each_entry_safe(e, h, n, &tun->flows[i], hash_link) {
unsigned long this_timer;
count++;
this_timer = e->updated + delay;
if (time_before_eq(this_timer, jiffies))
tun_flow_delete(tun, e);
else if (time_before(this_timer, next_timer))
next_timer = this_timer;
}
}
if (count)
mod_timer(&tun->flow_gc_timer, round_jiffies_up(next_timer));
spin_unlock_bh(&tun->lock);
}
static void tun_flow_update(struct tun_struct *tun, struct sk_buff *skb,
u16 queue_index)
{
struct hlist_head *head;
struct tun_flow_entry *e;
unsigned long delay = tun->ageing_time;
u32 rxhash = skb_get_rxhash(skb);
if (!rxhash)
return;
else
head = &tun->flows[tun_hashfn(rxhash)];
rcu_read_lock();
if (tun->numqueues == 1)
goto unlock;
e = tun_flow_find(head, rxhash);
if (likely(e)) {
/* TODO: keep queueing to old queue until it's empty? */
e->queue_index = queue_index;
e->updated = jiffies;
} else {
spin_lock_bh(&tun->lock);
if (!tun_flow_find(head, rxhash))
tun_flow_create(tun, head, rxhash, queue_index);
if (!timer_pending(&tun->flow_gc_timer))
mod_timer(&tun->flow_gc_timer,
round_jiffies_up(jiffies + delay));
spin_unlock_bh(&tun->lock);
}
unlock:
rcu_read_unlock();
}
/* We try to identify a flow through its rxhash first. The reason that /* We try to identify a flow through its rxhash first. The reason that
* we do not check rxq no. is becuase some cards(e.g 82599), chooses * we do not check rxq no. is becuase some cards(e.g 82599), chooses
* the rxq based on the txq where the last packet of the flow comes. As * the rxq based on the txq where the last packet of the flow comes. As
...@@ -175,6 +345,7 @@ struct tun_struct { ...@@ -175,6 +345,7 @@ struct tun_struct {
static u16 tun_select_queue(struct net_device *dev, struct sk_buff *skb) static u16 tun_select_queue(struct net_device *dev, struct sk_buff *skb)
{ {
struct tun_struct *tun = netdev_priv(dev); struct tun_struct *tun = netdev_priv(dev);
struct tun_flow_entry *e;
u32 txq = 0; u32 txq = 0;
u32 numqueues = 0; u32 numqueues = 0;
...@@ -183,8 +354,12 @@ static u16 tun_select_queue(struct net_device *dev, struct sk_buff *skb) ...@@ -183,8 +354,12 @@ static u16 tun_select_queue(struct net_device *dev, struct sk_buff *skb)
txq = skb_get_rxhash(skb); txq = skb_get_rxhash(skb);
if (txq) { if (txq) {
/* use multiply and shift instead of expensive divide */ e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq);
txq = ((u64)txq * numqueues) >> 32; if (e)
txq = e->queue_index;
else
/* use multiply and shift instead of expensive divide */
txq = ((u64)txq * numqueues) >> 32;
} else if (likely(skb_rx_queue_recorded(skb))) { } else if (likely(skb_rx_queue_recorded(skb))) {
txq = skb_get_rx_queue(skb); txq = skb_get_rx_queue(skb);
while (unlikely(txq >= numqueues)) while (unlikely(txq >= numqueues))
...@@ -234,6 +409,7 @@ static void __tun_detach(struct tun_file *tfile, bool clean) ...@@ -234,6 +409,7 @@ static void __tun_detach(struct tun_file *tfile, bool clean)
sock_put(&tfile->sk); sock_put(&tfile->sk);
synchronize_net(); synchronize_net();
tun_flow_delete_by_queue(tun, tun->numqueues + 1);
/* Drop read queue */ /* Drop read queue */
skb_queue_purge(&tfile->sk.sk_receive_queue); skb_queue_purge(&tfile->sk.sk_receive_queue);
tun_set_real_num_queues(tun); tun_set_real_num_queues(tun);
...@@ -631,6 +807,37 @@ static const struct net_device_ops tap_netdev_ops = { ...@@ -631,6 +807,37 @@ static const struct net_device_ops tap_netdev_ops = {
#endif #endif
}; };
static int tun_flow_init(struct tun_struct *tun)
{
int i;
tun->flow_cache = kmem_cache_create("tun_flow_cache",
sizeof(struct tun_flow_entry), 0, 0,
NULL);
if (!tun->flow_cache)
return -ENOMEM;
for (i = 0; i < TUN_NUM_FLOW_ENTRIES; i++)
INIT_HLIST_HEAD(&tun->flows[i]);
tun->ageing_time = TUN_FLOW_EXPIRE;
setup_timer(&tun->flow_gc_timer, tun_flow_cleanup, (unsigned long)tun);
mod_timer(&tun->flow_gc_timer,
round_jiffies_up(jiffies + tun->ageing_time));
return 0;
}
static void tun_flow_uninit(struct tun_struct *tun)
{
del_timer_sync(&tun->flow_gc_timer);
tun_flow_flush(tun);
/* Wait for completion of call_rcu()'s */
rcu_barrier();
kmem_cache_destroy(tun->flow_cache);
}
/* Initialize net device. */ /* Initialize net device. */
static void tun_net_init(struct net_device *dev) static void tun_net_init(struct net_device *dev)
{ {
...@@ -973,6 +1180,7 @@ static ssize_t tun_get_user(struct tun_struct *tun, struct tun_file *tfile, ...@@ -973,6 +1180,7 @@ static ssize_t tun_get_user(struct tun_struct *tun, struct tun_file *tfile,
tun->dev->stats.rx_packets++; tun->dev->stats.rx_packets++;
tun->dev->stats.rx_bytes += len; tun->dev->stats.rx_bytes += len;
tun_flow_update(tun, skb, tfile->queue_index);
return total_len; return total_len;
} }
...@@ -1150,6 +1358,14 @@ static ssize_t tun_chr_aio_read(struct kiocb *iocb, const struct iovec *iv, ...@@ -1150,6 +1358,14 @@ static ssize_t tun_chr_aio_read(struct kiocb *iocb, const struct iovec *iv,
return ret; return ret;
} }
static void tun_free_netdev(struct net_device *dev)
{
struct tun_struct *tun = netdev_priv(dev);
tun_flow_uninit(tun);
free_netdev(dev);
}
static void tun_setup(struct net_device *dev) static void tun_setup(struct net_device *dev)
{ {
struct tun_struct *tun = netdev_priv(dev); struct tun_struct *tun = netdev_priv(dev);
...@@ -1158,7 +1374,7 @@ static void tun_setup(struct net_device *dev) ...@@ -1158,7 +1374,7 @@ static void tun_setup(struct net_device *dev)
tun->group = INVALID_GID; tun->group = INVALID_GID;
dev->ethtool_ops = &tun_ethtool_ops; dev->ethtool_ops = &tun_ethtool_ops;
dev->destructor = free_netdev; dev->destructor = tun_free_netdev;
} }
/* Trivial set of netlink ops to allow deleting tun or tap /* Trivial set of netlink ops to allow deleting tun or tap
...@@ -1381,10 +1597,15 @@ static int tun_set_iff(struct net *net, struct file *file, struct ifreq *ifr) ...@@ -1381,10 +1597,15 @@ static int tun_set_iff(struct net *net, struct file *file, struct ifreq *ifr)
tun->filter_attached = false; tun->filter_attached = false;
tun->sndbuf = tfile->socket.sk->sk_sndbuf; tun->sndbuf = tfile->socket.sk->sk_sndbuf;
spin_lock_init(&tun->lock);
security_tun_dev_post_create(&tfile->sk); security_tun_dev_post_create(&tfile->sk);
tun_net_init(dev); tun_net_init(dev);
if (tun_flow_init(tun))
goto err_free_dev;
dev->hw_features = NETIF_F_SG | NETIF_F_FRAGLIST | dev->hw_features = NETIF_F_SG | NETIF_F_FRAGLIST |
TUN_USER_FEATURES; TUN_USER_FEATURES;
dev->features = dev->hw_features; dev->features = dev->hw_features;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment