Commit 1a77de09 authored by David S. Miller's avatar David S. Miller

Merge branch 'mptcp-optimizations'

Mat Martineau says:

====================
mptcp: A few optimizations

Here is a set of patches that we've accumulated and tested in the MPTCP
tree.

Patch 1 removes the MPTCP-level tx skb cache that added complexity but
did not provide a meaningful benefit.

Patch 2 uses the fast socket lock in more places.

Patch 3 improves handling of a data-ready flag.

Patch 4 deletes an unnecessary and racy connection state check.

Patch 5 adds a MIB counter for one type of invalid MPTCP header.

Patch 6 improves self test failure output.
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents 070258ef a4debc47
...@@ -25,6 +25,7 @@ static const struct snmp_mib mptcp_snmp_list[] = { ...@@ -25,6 +25,7 @@ static const struct snmp_mib mptcp_snmp_list[] = {
SNMP_MIB_ITEM("MPJoinAckHMacFailure", MPTCP_MIB_JOINACKMAC), SNMP_MIB_ITEM("MPJoinAckHMacFailure", MPTCP_MIB_JOINACKMAC),
SNMP_MIB_ITEM("DSSNotMatching", MPTCP_MIB_DSSNOMATCH), SNMP_MIB_ITEM("DSSNotMatching", MPTCP_MIB_DSSNOMATCH),
SNMP_MIB_ITEM("InfiniteMapRx", MPTCP_MIB_INFINITEMAPRX), SNMP_MIB_ITEM("InfiniteMapRx", MPTCP_MIB_INFINITEMAPRX),
SNMP_MIB_ITEM("DSSNoMatchTCP", MPTCP_MIB_DSSTCPMISMATCH),
SNMP_MIB_ITEM("DataCsumErr", MPTCP_MIB_DATACSUMERR), SNMP_MIB_ITEM("DataCsumErr", MPTCP_MIB_DATACSUMERR),
SNMP_MIB_ITEM("OFOQueueTail", MPTCP_MIB_OFOQUEUETAIL), SNMP_MIB_ITEM("OFOQueueTail", MPTCP_MIB_OFOQUEUETAIL),
SNMP_MIB_ITEM("OFOQueue", MPTCP_MIB_OFOQUEUE), SNMP_MIB_ITEM("OFOQueue", MPTCP_MIB_OFOQUEUE),
......
...@@ -18,6 +18,7 @@ enum linux_mptcp_mib_field { ...@@ -18,6 +18,7 @@ enum linux_mptcp_mib_field {
MPTCP_MIB_JOINACKMAC, /* HMAC was wrong on ACK + MP_JOIN */ MPTCP_MIB_JOINACKMAC, /* HMAC was wrong on ACK + MP_JOIN */
MPTCP_MIB_DSSNOMATCH, /* Received a new mapping that did not match the previous one */ MPTCP_MIB_DSSNOMATCH, /* Received a new mapping that did not match the previous one */
MPTCP_MIB_INFINITEMAPRX, /* Received an infinite mapping */ MPTCP_MIB_INFINITEMAPRX, /* Received an infinite mapping */
MPTCP_MIB_DSSTCPMISMATCH, /* DSS-mapping did not map with TCP's sequence numbers */
MPTCP_MIB_DATACSUMERR, /* The data checksum fail */ MPTCP_MIB_DATACSUMERR, /* The data checksum fail */
MPTCP_MIB_OFOQUEUETAIL, /* Segments inserted into OoO queue tail */ MPTCP_MIB_OFOQUEUETAIL, /* Segments inserted into OoO queue tail */
MPTCP_MIB_OFOQUEUE, /* Segments inserted into OoO queue */ MPTCP_MIB_OFOQUEUE, /* Segments inserted into OoO queue */
......
...@@ -540,6 +540,7 @@ void mptcp_pm_nl_addr_send_ack(struct mptcp_sock *msk) ...@@ -540,6 +540,7 @@ void mptcp_pm_nl_addr_send_ack(struct mptcp_sock *msk)
subflow = list_first_entry_or_null(&msk->conn_list, typeof(*subflow), node); subflow = list_first_entry_or_null(&msk->conn_list, typeof(*subflow), node);
if (subflow) { if (subflow) {
struct sock *ssk = mptcp_subflow_tcp_sock(subflow); struct sock *ssk = mptcp_subflow_tcp_sock(subflow);
bool slow;
spin_unlock_bh(&msk->pm.lock); spin_unlock_bh(&msk->pm.lock);
pr_debug("send ack for %s%s%s", pr_debug("send ack for %s%s%s",
...@@ -547,9 +548,9 @@ void mptcp_pm_nl_addr_send_ack(struct mptcp_sock *msk) ...@@ -547,9 +548,9 @@ void mptcp_pm_nl_addr_send_ack(struct mptcp_sock *msk)
mptcp_pm_should_add_signal_ipv6(msk) ? " [ipv6]" : "", mptcp_pm_should_add_signal_ipv6(msk) ? " [ipv6]" : "",
mptcp_pm_should_add_signal_port(msk) ? " [port]" : ""); mptcp_pm_should_add_signal_port(msk) ? " [port]" : "");
lock_sock(ssk); slow = lock_sock_fast(ssk);
tcp_send_ack(ssk); tcp_send_ack(ssk);
release_sock(ssk); unlock_sock_fast(ssk, slow);
spin_lock_bh(&msk->pm.lock); spin_lock_bh(&msk->pm.lock);
} }
} }
...@@ -566,6 +567,7 @@ int mptcp_pm_nl_mp_prio_send_ack(struct mptcp_sock *msk, ...@@ -566,6 +567,7 @@ int mptcp_pm_nl_mp_prio_send_ack(struct mptcp_sock *msk,
struct sock *ssk = mptcp_subflow_tcp_sock(subflow); struct sock *ssk = mptcp_subflow_tcp_sock(subflow);
struct sock *sk = (struct sock *)msk; struct sock *sk = (struct sock *)msk;
struct mptcp_addr_info local; struct mptcp_addr_info local;
bool slow;
local_address((struct sock_common *)ssk, &local); local_address((struct sock_common *)ssk, &local);
if (!addresses_equal(&local, addr, addr->port)) if (!addresses_equal(&local, addr, addr->port))
...@@ -578,9 +580,9 @@ int mptcp_pm_nl_mp_prio_send_ack(struct mptcp_sock *msk, ...@@ -578,9 +580,9 @@ int mptcp_pm_nl_mp_prio_send_ack(struct mptcp_sock *msk,
spin_unlock_bh(&msk->pm.lock); spin_unlock_bh(&msk->pm.lock);
pr_debug("send ack for mp_prio"); pr_debug("send ack for mp_prio");
lock_sock(ssk); slow = lock_sock_fast(ssk);
tcp_send_ack(ssk); tcp_send_ack(ssk);
release_sock(ssk); unlock_sock_fast(ssk, slow);
spin_lock_bh(&msk->pm.lock); spin_lock_bh(&msk->pm.lock);
return 0; return 0;
......
...@@ -433,23 +433,25 @@ static void mptcp_send_ack(struct mptcp_sock *msk) ...@@ -433,23 +433,25 @@ static void mptcp_send_ack(struct mptcp_sock *msk)
mptcp_for_each_subflow(msk, subflow) { mptcp_for_each_subflow(msk, subflow) {
struct sock *ssk = mptcp_subflow_tcp_sock(subflow); struct sock *ssk = mptcp_subflow_tcp_sock(subflow);
bool slow;
lock_sock(ssk); slow = lock_sock_fast(ssk);
if (tcp_can_send_ack(ssk)) if (tcp_can_send_ack(ssk))
tcp_send_ack(ssk); tcp_send_ack(ssk);
release_sock(ssk); unlock_sock_fast(ssk, slow);
} }
} }
static bool mptcp_subflow_cleanup_rbuf(struct sock *ssk) static bool mptcp_subflow_cleanup_rbuf(struct sock *ssk)
{ {
bool slow;
int ret; int ret;
lock_sock(ssk); slow = lock_sock_fast(ssk);
ret = tcp_can_send_ack(ssk); ret = tcp_can_send_ack(ssk);
if (ret) if (ret)
tcp_cleanup_rbuf(ssk, 1); tcp_cleanup_rbuf(ssk, 1);
release_sock(ssk); unlock_sock_fast(ssk, slow);
return ret; return ret;
} }
...@@ -684,9 +686,6 @@ static bool move_skbs_to_msk(struct mptcp_sock *msk, struct sock *ssk) ...@@ -684,9 +686,6 @@ static bool move_skbs_to_msk(struct mptcp_sock *msk, struct sock *ssk)
struct sock *sk = (struct sock *)msk; struct sock *sk = (struct sock *)msk;
unsigned int moved = 0; unsigned int moved = 0;
if (inet_sk_state_load(sk) == TCP_CLOSE)
return false;
__mptcp_move_skbs_from_subflow(msk, ssk, &moved); __mptcp_move_skbs_from_subflow(msk, ssk, &moved);
__mptcp_ofo_queue(msk); __mptcp_ofo_queue(msk);
if (unlikely(ssk->sk_err)) { if (unlikely(ssk->sk_err)) {
...@@ -902,22 +901,14 @@ static bool mptcp_frag_can_collapse_to(const struct mptcp_sock *msk, ...@@ -902,22 +901,14 @@ static bool mptcp_frag_can_collapse_to(const struct mptcp_sock *msk,
df->data_seq + df->data_len == msk->write_seq; df->data_seq + df->data_len == msk->write_seq;
} }
static int mptcp_wmem_with_overhead(struct sock *sk, int size) static int mptcp_wmem_with_overhead(int size)
{ {
struct mptcp_sock *msk = mptcp_sk(sk); return size + ((sizeof(struct mptcp_data_frag) * size) >> PAGE_SHIFT);
int ret, skbs;
ret = size + ((sizeof(struct mptcp_data_frag) * size) >> PAGE_SHIFT);
skbs = (msk->tx_pending_data + size) / msk->size_goal_cache;
if (skbs < msk->skb_tx_cache.qlen)
return ret;
return ret + (skbs - msk->skb_tx_cache.qlen) * SKB_TRUESIZE(MAX_TCP_HEADER);
} }
static void __mptcp_wmem_reserve(struct sock *sk, int size) static void __mptcp_wmem_reserve(struct sock *sk, int size)
{ {
int amount = mptcp_wmem_with_overhead(sk, size); int amount = mptcp_wmem_with_overhead(size);
struct mptcp_sock *msk = mptcp_sk(sk); struct mptcp_sock *msk = mptcp_sk(sk);
WARN_ON_ONCE(msk->wmem_reserved); WARN_ON_ONCE(msk->wmem_reserved);
...@@ -1212,49 +1203,8 @@ static struct sk_buff *__mptcp_do_alloc_tx_skb(struct sock *sk, gfp_t gfp) ...@@ -1212,49 +1203,8 @@ static struct sk_buff *__mptcp_do_alloc_tx_skb(struct sock *sk, gfp_t gfp)
return NULL; return NULL;
} }
static bool mptcp_tx_cache_refill(struct sock *sk, int size,
struct sk_buff_head *skbs, int *total_ts)
{
struct mptcp_sock *msk = mptcp_sk(sk);
struct sk_buff *skb;
int space_needed;
if (unlikely(tcp_under_memory_pressure(sk))) {
mptcp_mem_reclaim_partial(sk);
/* under pressure pre-allocate at most a single skb */
if (msk->skb_tx_cache.qlen)
return true;
space_needed = msk->size_goal_cache;
} else {
space_needed = msk->tx_pending_data + size -
msk->skb_tx_cache.qlen * msk->size_goal_cache;
}
while (space_needed > 0) {
skb = __mptcp_do_alloc_tx_skb(sk, sk->sk_allocation);
if (unlikely(!skb)) {
/* under memory pressure, try to pass the caller a
* single skb to allow forward progress
*/
while (skbs->qlen > 1) {
skb = __skb_dequeue_tail(skbs);
*total_ts -= skb->truesize;
__kfree_skb(skb);
}
return skbs->qlen > 0;
}
*total_ts += skb->truesize;
__skb_queue_tail(skbs, skb);
space_needed -= msk->size_goal_cache;
}
return true;
}
static bool __mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp) static bool __mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp)
{ {
struct mptcp_sock *msk = mptcp_sk(sk);
struct sk_buff *skb; struct sk_buff *skb;
if (ssk->sk_tx_skb_cache) { if (ssk->sk_tx_skb_cache) {
...@@ -1265,22 +1215,6 @@ static bool __mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp) ...@@ -1265,22 +1215,6 @@ static bool __mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp)
return true; return true;
} }
skb = skb_peek(&msk->skb_tx_cache);
if (skb) {
if (likely(sk_wmem_schedule(ssk, skb->truesize))) {
skb = __skb_dequeue(&msk->skb_tx_cache);
if (WARN_ON_ONCE(!skb))
return false;
mptcp_wmem_uncharge(sk, skb->truesize);
ssk->sk_tx_skb_cache = skb;
return true;
}
/* over memory limit, no point to try to allocate a new skb */
return false;
}
skb = __mptcp_do_alloc_tx_skb(sk, gfp); skb = __mptcp_do_alloc_tx_skb(sk, gfp);
if (!skb) if (!skb)
return false; return false;
...@@ -1296,7 +1230,6 @@ static bool __mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp) ...@@ -1296,7 +1230,6 @@ static bool __mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp)
static bool mptcp_must_reclaim_memory(struct sock *sk, struct sock *ssk) static bool mptcp_must_reclaim_memory(struct sock *sk, struct sock *ssk)
{ {
return !ssk->sk_tx_skb_cache && return !ssk->sk_tx_skb_cache &&
!skb_peek(&mptcp_sk(sk)->skb_tx_cache) &&
tcp_under_memory_pressure(sk); tcp_under_memory_pressure(sk);
} }
...@@ -1339,7 +1272,6 @@ static int mptcp_sendmsg_frag(struct sock *sk, struct sock *ssk, ...@@ -1339,7 +1272,6 @@ static int mptcp_sendmsg_frag(struct sock *sk, struct sock *ssk,
/* compute send limit */ /* compute send limit */
info->mss_now = tcp_send_mss(ssk, &info->size_goal, info->flags); info->mss_now = tcp_send_mss(ssk, &info->size_goal, info->flags);
avail_size = info->size_goal; avail_size = info->size_goal;
msk->size_goal_cache = info->size_goal;
skb = tcp_write_queue_tail(ssk); skb = tcp_write_queue_tail(ssk);
if (skb) { if (skb) {
/* Limit the write to the size available in the /* Limit the write to the size available in the
...@@ -1688,7 +1620,6 @@ static int mptcp_sendmsg(struct sock *sk, struct msghdr *msg, size_t len) ...@@ -1688,7 +1620,6 @@ static int mptcp_sendmsg(struct sock *sk, struct msghdr *msg, size_t len)
while (msg_data_left(msg)) { while (msg_data_left(msg)) {
int total_ts, frag_truesize = 0; int total_ts, frag_truesize = 0;
struct mptcp_data_frag *dfrag; struct mptcp_data_frag *dfrag;
struct sk_buff_head skbs;
bool dfrag_collapsed; bool dfrag_collapsed;
size_t psize, offset; size_t psize, offset;
...@@ -1721,16 +1652,10 @@ static int mptcp_sendmsg(struct sock *sk, struct msghdr *msg, size_t len) ...@@ -1721,16 +1652,10 @@ static int mptcp_sendmsg(struct sock *sk, struct msghdr *msg, size_t len)
psize = pfrag->size - offset; psize = pfrag->size - offset;
psize = min_t(size_t, psize, msg_data_left(msg)); psize = min_t(size_t, psize, msg_data_left(msg));
total_ts = psize + frag_truesize; total_ts = psize + frag_truesize;
__skb_queue_head_init(&skbs);
if (!mptcp_tx_cache_refill(sk, psize, &skbs, &total_ts))
goto wait_for_memory;
if (!mptcp_wmem_alloc(sk, total_ts)) { if (!mptcp_wmem_alloc(sk, total_ts))
__skb_queue_purge(&skbs);
goto wait_for_memory; goto wait_for_memory;
}
skb_queue_splice_tail(&skbs, &msk->skb_tx_cache);
if (copy_page_from_iter(dfrag->page, offset, psize, if (copy_page_from_iter(dfrag->page, offset, psize,
&msg->msg_iter) != psize) { &msg->msg_iter) != psize) {
mptcp_wmem_uncharge(sk, psize + frag_truesize); mptcp_wmem_uncharge(sk, psize + frag_truesize);
...@@ -1787,7 +1712,7 @@ static void mptcp_wait_data(struct sock *sk, long *timeo) ...@@ -1787,7 +1712,7 @@ static void mptcp_wait_data(struct sock *sk, long *timeo)
sk_set_bit(SOCKWQ_ASYNC_WAITDATA, sk); sk_set_bit(SOCKWQ_ASYNC_WAITDATA, sk);
sk_wait_event(sk, timeo, sk_wait_event(sk, timeo,
test_and_clear_bit(MPTCP_DATA_READY, &msk->flags), &wait); test_bit(MPTCP_DATA_READY, &msk->flags), &wait);
sk_clear_bit(SOCKWQ_ASYNC_WAITDATA, sk); sk_clear_bit(SOCKWQ_ASYNC_WAITDATA, sk);
remove_wait_queue(sk_sleep(sk), &wait); remove_wait_queue(sk_sleep(sk), &wait);
...@@ -2111,10 +2036,8 @@ static int mptcp_recvmsg(struct sock *sk, struct msghdr *msg, size_t len, ...@@ -2111,10 +2036,8 @@ static int mptcp_recvmsg(struct sock *sk, struct msghdr *msg, size_t len,
*/ */
if (unlikely(__mptcp_move_skbs(msk))) if (unlikely(__mptcp_move_skbs(msk)))
set_bit(MPTCP_DATA_READY, &msk->flags); set_bit(MPTCP_DATA_READY, &msk->flags);
} else if (unlikely(!test_bit(MPTCP_DATA_READY, &msk->flags))) {
/* data to read but mptcp_wait_data() cleared DATA_READY */
set_bit(MPTCP_DATA_READY, &msk->flags);
} }
out_err: out_err:
if (cmsg_flags && copied >= 0) { if (cmsg_flags && copied >= 0) {
if (cmsg_flags & MPTCP_CMSG_TS) if (cmsg_flags & MPTCP_CMSG_TS)
...@@ -2326,13 +2249,14 @@ static void mptcp_check_fastclose(struct mptcp_sock *msk) ...@@ -2326,13 +2249,14 @@ static void mptcp_check_fastclose(struct mptcp_sock *msk)
list_for_each_entry_safe(subflow, tmp, &msk->conn_list, node) { list_for_each_entry_safe(subflow, tmp, &msk->conn_list, node) {
struct sock *tcp_sk = mptcp_subflow_tcp_sock(subflow); struct sock *tcp_sk = mptcp_subflow_tcp_sock(subflow);
bool slow;
lock_sock(tcp_sk); slow = lock_sock_fast(tcp_sk);
if (tcp_sk->sk_state != TCP_CLOSE) { if (tcp_sk->sk_state != TCP_CLOSE) {
tcp_send_active_reset(tcp_sk, GFP_ATOMIC); tcp_send_active_reset(tcp_sk, GFP_ATOMIC);
tcp_set_state(tcp_sk, TCP_CLOSE); tcp_set_state(tcp_sk, TCP_CLOSE);
} }
release_sock(tcp_sk); unlock_sock_fast(tcp_sk, slow);
} }
inet_sk_state_store(sk, TCP_CLOSE); inet_sk_state_store(sk, TCP_CLOSE);
...@@ -2462,13 +2386,11 @@ static int __mptcp_init_sock(struct sock *sk) ...@@ -2462,13 +2386,11 @@ static int __mptcp_init_sock(struct sock *sk)
INIT_LIST_HEAD(&msk->rtx_queue); INIT_LIST_HEAD(&msk->rtx_queue);
INIT_WORK(&msk->work, mptcp_worker); INIT_WORK(&msk->work, mptcp_worker);
__skb_queue_head_init(&msk->receive_queue); __skb_queue_head_init(&msk->receive_queue);
__skb_queue_head_init(&msk->skb_tx_cache);
msk->out_of_order_queue = RB_ROOT; msk->out_of_order_queue = RB_ROOT;
msk->first_pending = NULL; msk->first_pending = NULL;
msk->wmem_reserved = 0; msk->wmem_reserved = 0;
msk->rmem_released = 0; msk->rmem_released = 0;
msk->tx_pending_data = 0; msk->tx_pending_data = 0;
msk->size_goal_cache = TCP_BASE_MSS;
msk->ack_hint = NULL; msk->ack_hint = NULL;
msk->first = NULL; msk->first = NULL;
...@@ -2525,15 +2447,10 @@ static void __mptcp_clear_xmit(struct sock *sk) ...@@ -2525,15 +2447,10 @@ static void __mptcp_clear_xmit(struct sock *sk)
{ {
struct mptcp_sock *msk = mptcp_sk(sk); struct mptcp_sock *msk = mptcp_sk(sk);
struct mptcp_data_frag *dtmp, *dfrag; struct mptcp_data_frag *dtmp, *dfrag;
struct sk_buff *skb;
WRITE_ONCE(msk->first_pending, NULL); WRITE_ONCE(msk->first_pending, NULL);
list_for_each_entry_safe(dfrag, dtmp, &msk->rtx_queue, list) list_for_each_entry_safe(dfrag, dtmp, &msk->rtx_queue, list)
dfrag_clear(sk, dfrag); dfrag_clear(sk, dfrag);
while ((skb = __skb_dequeue(&msk->skb_tx_cache)) != NULL) {
sk->sk_forward_alloc += skb->truesize;
kfree_skb(skb);
}
} }
static void mptcp_cancel_work(struct sock *sk) static void mptcp_cancel_work(struct sock *sk)
......
...@@ -245,9 +245,7 @@ struct mptcp_sock { ...@@ -245,9 +245,7 @@ struct mptcp_sock {
struct sk_buff *ooo_last_skb; struct sk_buff *ooo_last_skb;
struct rb_root out_of_order_queue; struct rb_root out_of_order_queue;
struct sk_buff_head receive_queue; struct sk_buff_head receive_queue;
struct sk_buff_head skb_tx_cache; /* this is wmem accounted */
int tx_pending_data; int tx_pending_data;
int size_goal_cache;
struct list_head conn_list; struct list_head conn_list;
struct list_head rtx_queue; struct list_head rtx_queue;
struct mptcp_data_frag *first_pending; struct mptcp_data_frag *first_pending;
......
...@@ -1046,8 +1046,10 @@ static enum mapping_status get_mapping_status(struct sock *ssk, ...@@ -1046,8 +1046,10 @@ static enum mapping_status get_mapping_status(struct sock *ssk,
/* we revalidate valid mapping on new skb, because we must ensure /* we revalidate valid mapping on new skb, because we must ensure
* the current skb is completely covered by the available mapping * the current skb is completely covered by the available mapping
*/ */
if (!validate_mapping(ssk, skb)) if (!validate_mapping(ssk, skb)) {
MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DSSTCPMISMATCH);
return MAPPING_INVALID; return MAPPING_INVALID;
}
skb_ext_del(skb, SKB_EXT_MPTCP); skb_ext_del(skb, SKB_EXT_MPTCP);
......
...@@ -680,6 +680,25 @@ run_tests_peekmode() ...@@ -680,6 +680,25 @@ run_tests_peekmode()
run_tests_lo "$ns1" "$ns1" dead:beef:1::1 1 "-P ${peekmode}" run_tests_lo "$ns1" "$ns1" dead:beef:1::1 1 "-P ${peekmode}"
} }
display_time()
{
time_end=$(date +%s)
time_run=$((time_end-time_start))
echo "Time: ${time_run} seconds"
}
stop_if_error()
{
local msg="$1"
if [ ${ret} -ne 0 ]; then
echo "FAIL: ${msg}" 1>&2
display_time
exit ${ret}
fi
}
make_file "$cin" "client" make_file "$cin" "client"
make_file "$sin" "server" make_file "$sin" "server"
...@@ -687,6 +706,8 @@ check_mptcp_disabled ...@@ -687,6 +706,8 @@ check_mptcp_disabled
check_mptcp_ulp_setsockopt check_mptcp_ulp_setsockopt
stop_if_error "The kernel configuration is not valid for MPTCP"
echo "INFO: validating network environment with pings" echo "INFO: validating network environment with pings"
for sender in "$ns1" "$ns2" "$ns3" "$ns4";do for sender in "$ns1" "$ns2" "$ns3" "$ns4";do
do_ping "$ns1" $sender 10.0.1.1 do_ping "$ns1" $sender 10.0.1.1
...@@ -706,6 +727,8 @@ for sender in "$ns1" "$ns2" "$ns3" "$ns4";do ...@@ -706,6 +727,8 @@ for sender in "$ns1" "$ns2" "$ns3" "$ns4";do
do_ping "$ns4" $sender dead:beef:3::1 do_ping "$ns4" $sender dead:beef:3::1
done done
stop_if_error "Could not even run ping tests"
[ -n "$tc_loss" ] && tc -net "$ns2" qdisc add dev ns2eth3 root netem loss random $tc_loss delay ${tc_delay}ms [ -n "$tc_loss" ] && tc -net "$ns2" qdisc add dev ns2eth3 root netem loss random $tc_loss delay ${tc_delay}ms
echo -n "INFO: Using loss of $tc_loss " echo -n "INFO: Using loss of $tc_loss "
test "$tc_delay" -gt 0 && echo -n "delay $tc_delay ms " test "$tc_delay" -gt 0 && echo -n "delay $tc_delay ms "
...@@ -733,18 +756,13 @@ echo "on ns3eth4" ...@@ -733,18 +756,13 @@ echo "on ns3eth4"
tc -net "$ns3" qdisc add dev ns3eth4 root netem delay ${reorder_delay}ms $tc_reorder tc -net "$ns3" qdisc add dev ns3eth4 root netem delay ${reorder_delay}ms $tc_reorder
for sender in $ns1 $ns2 $ns3 $ns4;do run_tests_lo "$ns1" "$ns1" 10.0.1.1 1
run_tests_lo "$ns1" "$sender" 10.0.1.1 1 stop_if_error "Could not even run loopback test"
if [ $ret -ne 0 ] ;then
echo "FAIL: Could not even run loopback test" 1>&2 run_tests_lo "$ns1" "$ns1" dead:beef:1::1 1
exit $ret stop_if_error "Could not even run loopback v6 test"
fi
run_tests_lo "$ns1" $sender dead:beef:1::1 1
if [ $ret -ne 0 ] ;then
echo "FAIL: Could not even run loopback v6 test" 2>&1
exit $ret
fi
for sender in $ns1 $ns2 $ns3 $ns4;do
# ns1<->ns2 is not subject to reordering/tc delays. Use it to test # ns1<->ns2 is not subject to reordering/tc delays. Use it to test
# mptcp syncookie support. # mptcp syncookie support.
if [ $sender = $ns1 ]; then if [ $sender = $ns1 ]; then
...@@ -753,6 +771,9 @@ for sender in $ns1 $ns2 $ns3 $ns4;do ...@@ -753,6 +771,9 @@ for sender in $ns1 $ns2 $ns3 $ns4;do
ip netns exec "$ns2" sysctl -q net.ipv4.tcp_syncookies=1 ip netns exec "$ns2" sysctl -q net.ipv4.tcp_syncookies=1
fi fi
run_tests "$ns1" $sender 10.0.1.1
run_tests "$ns1" $sender dead:beef:1::1
run_tests "$ns2" $sender 10.0.1.2 run_tests "$ns2" $sender 10.0.1.2
run_tests "$ns2" $sender dead:beef:1::2 run_tests "$ns2" $sender dead:beef:1::2
run_tests "$ns2" $sender 10.0.2.1 run_tests "$ns2" $sender 10.0.2.1
...@@ -765,14 +786,13 @@ for sender in $ns1 $ns2 $ns3 $ns4;do ...@@ -765,14 +786,13 @@ for sender in $ns1 $ns2 $ns3 $ns4;do
run_tests "$ns4" $sender 10.0.3.1 run_tests "$ns4" $sender 10.0.3.1
run_tests "$ns4" $sender dead:beef:3::1 run_tests "$ns4" $sender dead:beef:3::1
stop_if_error "Tests with $sender as a sender have failed"
done done
run_tests_peekmode "saveWithPeek" run_tests_peekmode "saveWithPeek"
run_tests_peekmode "saveAfterPeek" run_tests_peekmode "saveAfterPeek"
stop_if_error "Tests with peek mode have failed"
time_end=$(date +%s) display_time
time_run=$((time_end-time_start))
echo "Time: ${time_run} seconds"
exit $ret exit $ret
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment