1. 17 Sep, 2018 19 commits
  2. 16 Sep, 2018 6 commits
  3. 14 Sep, 2018 4 commits
  4. 13 Sep, 2018 11 commits
    • Kees Cook's avatar
      net/ibm/emac: Remove VLA usage · ee4fccbe
      Kees Cook authored
      In the quest to remove all stack VLA usage from the kernel[1], this
      removes the VLA used for the emac xaht registers size. Since the size
      of registers can only ever be 4 or 8, as detected in emac_init_config(),
      the max can be hardcoded and a runtime test added for robustness.
      
      [1] https://lkml.kernel.org/r/CA+55aFzCG-zNmZwX4A2FQpadafLfEzK6CC=qPXydAacU1RqZWA@mail.gmail.com
      
      Cc: "David S. Miller" <davem@davemloft.net>
      Cc: Christian Lamparter <chunkeey@gmail.com>
      Cc: Ivan Mikhaylov <ivan@de.ibm.com>
      Cc: netdev@vger.kernel.org
      Co-developed-by: default avatarBenjamin Herrenschmidt <benh@kernel.crashing.org>
      Signed-off-by: default avatarKees Cook <keescook@chromium.org>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      ee4fccbe
    • Gustavo A. R. Silva's avatar
      pktgen: Fix fall-through annotation · f91845da
      Gustavo A. R. Silva authored
      Replace "fallthru" with a proper "fall through" annotation.
      
      This fix is part of the ongoing efforts to enabling
      -Wimplicit-fallthrough
      Signed-off-by: default avatarGustavo A. R. Silva <gustavo@embeddedor.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      f91845da
    • Gustavo A. R. Silva's avatar
      tg3: Fix fall-through annotations · 310fc051
      Gustavo A. R. Silva authored
      Replace "fallthru" with a proper "fall through" annotation.
      
      This fix is part of the ongoing efforts to enabling
      -Wimplicit-fallthrough
      Signed-off-by: default avatarGustavo A. R. Silva <gustavo@embeddedor.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      310fc051
    • Toke Høiland-Jørgensen's avatar
      gso_segment: Reset skb->mac_len after modifying network header · 50c12f74
      Toke Høiland-Jørgensen authored
      When splitting a GSO segment that consists of encapsulated packets, the
      skb->mac_len of the segments can end up being set wrong, causing packet
      drops in particular when using act_mirred and ifb interfaces in
      combination with a qdisc that splits GSO packets.
      
      This happens because at the time skb_segment() is called, network_header
      will point to the inner header, throwing off the calculation in
      skb_reset_mac_len(). The network_header is subsequently adjust by the
      outer IP gso_segment handlers, but they don't set the mac_len.
      
      Fix this by adding skb_reset_mac_len() calls to both the IPv4 and IPv6
      gso_segment handlers, after they modify the network_header.
      
      Many thanks to Eric Dumazet for his help in identifying the cause of
      the bug.
      Acked-by: default avatarDave Taht <dave.taht@gmail.com>
      Reviewed-by: default avatarEric Dumazet <edumazet@google.com>
      Signed-off-by: default avatarToke Høiland-Jørgensen <toke@toke.dk>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      50c12f74
    • YueHaibing's avatar
      vxlan: Remove duplicated include from vxlan.h · 293681f1
      YueHaibing authored
      Remove duplicated include.
      Signed-off-by: default avatarYueHaibing <yuehaibing@huawei.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      293681f1
    • Florian Fainelli's avatar
      net: dsa: b53: Do not fail when IRQ are not initialized · b2ddc48a
      Florian Fainelli authored
      When the Device Tree is not providing the per-port interrupts, do not fail
      during b53_srab_irq_enable() but instead bail out gracefully. The SRAB driver
      is used on the BCM5301X (Northstar) platforms which do not yet have the SRAB
      interrupts wired up.
      
      Fixes: 16994374 ("net: dsa: b53: Make SRAB driver manage port interrupts")
      Signed-off-by: default avatarFlorian Fainelli <f.fainelli@gmail.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      b2ddc48a
    • David S. Miller's avatar
      Merge branch 'vhost_net-TX-batching' · 8bb83b78
      David S. Miller authored
      Jason Wang says:
      
      ====================
      vhost_net TX batching
      
      This series tries to batch submitting packets to underlayer socket
      through msg_control during sendmsg(). This is done by:
      
      1) Doing userspace copy inside vhost_net
      2) Build XDP buff
      3) Batch at most 64 (VHOST_NET_BATCH) XDP buffs and submit them once
         through msg_control during sendmsg().
      4) Underlayer sockets can use XDP buffs directly when XDP is enalbed,
         or build skb based on XDP buff.
      
      For the packet that can not be built easily with XDP or for the case
      that batch submission is hard (e.g sndbuf is limited). We will go for
      the previous slow path, passing iov iterator to underlayer socket
      through sendmsg() once per packet.
      
      This can help to improve cache utilization and avoid lots of indirect
      calls with sendmsg(). It can also co-operate with the batching support
      of the underlayer sockets (e.g the case of XDP redirection through
      maps).
      
      Testpmd(txonly) in guest shows obvious improvements:
      
      Test                /+pps%
      XDP_DROP on TAP     /+44.8%
      XDP_REDIRECT on TAP /+29%
      macvtap (skb)       /+26%
      
      Netperf TCP_STREAM TX from guest shows obvious improvements on small
      packet:
      
          size/session/+thu%/+normalize%
             64/     1/   +2%/    0%
             64/     2/   +3%/   +1%
             64/     4/   +7%/   +5%
             64/     8/   +8%/   +6%
            256/     1/   +3%/    0%
            256/     2/  +10%/   +7%
            256/     4/  +26%/  +22%
            256/     8/  +27%/  +23%
            512/     1/   +3%/   +2%
            512/     2/  +19%/  +14%
            512/     4/  +43%/  +40%
            512/     8/  +45%/  +41%
           1024/     1/   +4%/    0%
           1024/     2/  +27%/  +21%
           1024/     4/  +38%/  +73%
           1024/     8/  +15%/  +24%
           2048/     1/  +10%/   +7%
           2048/     2/  +16%/  +12%
           2048/     4/    0%/   +2%
           2048/     8/    0%/   +2%
           4096/     1/  +36%/  +60%
           4096/     2/  -11%/  -26%
           4096/     4/    0%/  +14%
           4096/     8/    0%/   +4%
          16384/     1/   -1%/   +5%
          16384/     2/    0%/   +2%
          16384/     4/    0%/   -3%
          16384/     8/    0%/   +4%
          65535/     1/    0%/  +10%
          65535/     2/    0%/   +8%
          65535/     4/    0%/   +1%
          65535/     8/    0%/   +3%
      
      Please review.
      ====================
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      8bb83b78
    • Jason Wang's avatar
      vhost_net: batch submitting XDP buffers to underlayer sockets · 0a0be13b
      Jason Wang authored
      This patch implements XDP batching for vhost_net. The idea is first to
      try to do userspace copy and build XDP buff directly in vhost. Instead
      of submitting the packet immediately, vhost_net will batch them in an
      array and submit every 64 (VHOST_NET_BATCH) packets to the under layer
      sockets through msg_control of sendmsg().
      
      When XDP is enabled on the TUN/TAP, TUN/TAP can process XDP inside a
      loop without caring GUP thus it can do batch map flushing. When XDP is
      not enabled or not supported, the underlayer socket need to build skb
      and pass it to network core. The batched packet submission allows us
      to do batching like netif_receive_skb_list() in the future.
      
      This saves lots of indirect calls for better cache utilization. For
      the case that we can't so batching e.g when sndbuf is limited or
      packet size is too large, we will go for usual one packet per
      sendmsg() way.
      
      Doing testpmd on various setups gives us:
      
      Test                /+pps%
      XDP_DROP on TAP     /+44.8%
      XDP_REDIRECT on TAP /+29%
      macvtap (skb)       /+26%
      
      Netperf tests shows obvious improvements for small packet transmission:
      
      size/session/+thu%/+normalize%
         64/     1/   +2%/    0%
         64/     2/   +3%/   +1%
         64/     4/   +7%/   +5%
         64/     8/   +8%/   +6%
        256/     1/   +3%/    0%
        256/     2/  +10%/   +7%
        256/     4/  +26%/  +22%
        256/     8/  +27%/  +23%
        512/     1/   +3%/   +2%
        512/     2/  +19%/  +14%
        512/     4/  +43%/  +40%
        512/     8/  +45%/  +41%
       1024/     1/   +4%/    0%
       1024/     2/  +27%/  +21%
       1024/     4/  +38%/  +73%
       1024/     8/  +15%/  +24%
       2048/     1/  +10%/   +7%
       2048/     2/  +16%/  +12%
       2048/     4/    0%/   +2%
       2048/     8/    0%/   +2%
       4096/     1/  +36%/  +60%
       4096/     2/  -11%/  -26%
       4096/     4/    0%/  +14%
       4096/     8/    0%/   +4%
      16384/     1/   -1%/   +5%
      16384/     2/    0%/   +2%
      16384/     4/    0%/   -3%
      16384/     8/    0%/   +4%
      65535/     1/    0%/  +10%
      65535/     2/    0%/   +8%
      65535/     4/    0%/   +1%
      65535/     8/    0%/   +3%
      Signed-off-by: default avatarJason Wang <jasowang@redhat.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      0a0be13b
    • Jason Wang's avatar
      tap: accept an array of XDP buffs through sendmsg() · 0efac277
      Jason Wang authored
      This patch implement TUN_MSG_PTR msg_control type. This type allows
      the caller to pass an array of XDP buffs to tuntap through ptr field
      of the tun_msg_control. Tap will build skb through those XDP buffers.
      
      This will avoid lots of indirect calls thus improves the icache
      utilization and allows to do XDP batched flushing when doing XDP
      redirection.
      Signed-off-by: default avatarJason Wang <jasowang@redhat.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      0efac277
    • Jason Wang's avatar
      tuntap: accept an array of XDP buffs through sendmsg() · 043d222f
      Jason Wang authored
      This patch implement TUN_MSG_PTR msg_control type. This type allows
      the caller to pass an array of XDP buffs to tuntap through ptr field
      of the tun_msg_control. If an XDP program is attached, tuntap can run
      XDP program directly. If not, tuntap will build skb and do a fast
      receiving since part of the work has been done by vhost_net.
      
      This will avoid lots of indirect calls thus improves the icache
      utilization and allows to do XDP batched flushing when doing XDP
      redirection.
      Signed-off-by: default avatarJason Wang <jasowang@redhat.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      043d222f
    • Jason Wang's avatar
      tun: switch to new type of msg_control · fe8dd45b
      Jason Wang authored
      This patch introduces to a new tun/tap specific msg_control:
      
      #define TUN_MSG_UBUF 1
      #define TUN_MSG_PTR  2
      struct tun_msg_ctl {
             int type;
             void *ptr;
      };
      
      This allows us to pass different kinds of msg_control through
      sendmsg(). The first supported type is ubuf (TUN_MSG_UBUF) which will
      be used by the existed vhost_net zerocopy code. The second is XDP
      buff, which allows vhost_net to pass XDP buff to TUN. This could be
      used to implement accepting an array of XDP buffs from vhost_net in
      the following patches.
      Signed-off-by: default avatarJason Wang <jasowang@redhat.com>
      Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
      fe8dd45b