53ff387156
For the TCP throughput tests, we use iperf3's -O "omit" option which ignores results for the given time at the beginning of the test. Currently we calculate this as 1/6th of the test measurement time. The purpose of -O, however, is to skip over the TCP slow start period, which in no way depends on the overall length of the test. The slow start time is roughly speaking log_2 ( max_window_size / MSS ) * round_trip_time These factors all vary between tests and machines we're running on, but we can estimate some reasonable bounds for them: * The maximum window size is bounded by the buffer sizes at each end, which shouldn't exceed 16MiB * The mss varies with the MTU we use, but the smallest we use in tests is ~256 bytes * Round trip time will vary with the system, but with these essentially local transfers it will typically be well under 1ms (on my laptop it is closer to 0.03ms) That gives a worst case slow start time of about 16ms. Setting an omit time of 0.1s uniformly is therefore more than enough, and substantially smaller than what we calculate now for the default case (10s / 6 ~= 1.7s). This reduces total time for the standard benchmark run by around 30s. Signed-off-by: David Gibson <david@gibson.dropbear.id.au> Signed-off-by: Stefano Brivio <sbrivio@redhat.com>
207 lines
5.5 KiB
Text
207 lines
5.5 KiB
Text
# SPDX-License-Identifier: GPL-2.0-or-later
|
|
#
|
|
# PASST - Plug A Simple Socket Transport
|
|
# for qemu/UNIX domain socket mode
|
|
#
|
|
# PASTA - Pack A Subtle Tap Abstraction
|
|
# for network namespace/tap device mode
|
|
#
|
|
# test/perf/passt_tcp - Check TCP performance in passt mode
|
|
#
|
|
# Copyright (c) 2021 Red Hat GmbH
|
|
# Author: Stefano Brivio <sbrivio@redhat.com>
|
|
|
|
gtools /sbin/sysctl ip jq nproc seq sleep iperf3 tcp_rr tcp_crr # From neper
|
|
nstools /sbin/sysctl ip jq nproc seq sleep iperf3 tcp_rr tcp_crr
|
|
htools bc head sed seq
|
|
|
|
test passt: throughput and latency
|
|
|
|
guest /sbin/sysctl -w net.core.rmem_max=536870912
|
|
guest /sbin/sysctl -w net.core.wmem_max=536870912
|
|
guest /sbin/sysctl -w net.core.rmem_default=33554432
|
|
guest /sbin/sysctl -w net.core.wmem_default=33554432
|
|
guest /sbin/sysctl -w net.ipv4.tcp_rmem="4096 131072 268435456"
|
|
guest /sbin/sysctl -w net.ipv4.tcp_wmem="4096 131072 268435456"
|
|
guest /sbin/sysctl -w net.ipv4.tcp_timestamps=0
|
|
|
|
ns /sbin/sysctl -w net.ipv4.tcp_rmem="4096 524288 134217728"
|
|
ns /sbin/sysctl -w net.ipv4.tcp_wmem="4096 524288 134217728"
|
|
ns /sbin/sysctl -w net.ipv4.tcp_timestamps=0
|
|
|
|
gout GW ip -j -4 route show|jq -rM '.[] | select(.dst == "default").gateway'
|
|
gout GW6 ip -j -6 route show|jq -rM '.[] | select(.dst == "default").gateway'
|
|
gout IFNAME ip -j link show | jq -rM '.[] | select(.link_type == "ether").ifname'
|
|
|
|
hout FREQ_PROCFS (echo "scale=1"; sed -n 's/cpu MHz.*: \([0-9]*\)\..*$/(\1+10^2\/2)\/10^3/p' /proc/cpuinfo) | bc -l | head -n1
|
|
hout FREQ_CPUFREQ (echo "scale=1"; printf '( %i + 10^5 / 2 ) / 10^6\n' $(cat /sys/devices/system/cpu/cpu0/cpufreq/scaling_max_freq) ) | bc -l
|
|
hout FREQ [ -n "__FREQ_CPUFREQ__" ] && echo __FREQ_CPUFREQ__ || echo __FREQ_PROCFS__
|
|
|
|
set THREADS 1
|
|
set STREAMS 8
|
|
set TIME 10
|
|
set OMIT 0.1
|
|
set OPTS -Z -P __STREAMS__ -l 1M -O__OMIT__
|
|
|
|
info Throughput in Gbps, latency in µs, one thread at __FREQ__ GHz, __STREAMS__ streams
|
|
report passt tcp __THREADS__ __FREQ__
|
|
|
|
th MTU 256B 576B 1280B 1500B 9000B 65520B
|
|
|
|
|
|
tr TCP throughput over IPv6: guest to host
|
|
iperf3s ns 100${i}2 __THREADS__
|
|
|
|
bw -
|
|
bw -
|
|
guest ip link set dev __IFNAME__ mtu 1280
|
|
iperf3 BW guest __GW6__%__IFNAME__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 4M
|
|
bw __BW__ 1.2 1.5
|
|
guest ip link set dev __IFNAME__ mtu 1500
|
|
iperf3 BW guest __GW6__%__IFNAME__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 4M
|
|
bw __BW__ 1.6 1.8
|
|
guest ip link set dev __IFNAME__ mtu 9000
|
|
iperf3 BW guest __GW6__%__IFNAME__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 8M
|
|
bw __BW__ 4.0 5.0
|
|
guest ip link set dev __IFNAME__ mtu 65520
|
|
iperf3 BW guest __GW6__%__IFNAME__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 16M
|
|
bw __BW__ 7.0 8.0
|
|
|
|
iperf3k ns
|
|
|
|
tl TCP RR latency over IPv6: guest to host
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
nsb tcp_rr --nolog -6
|
|
gout LAT tcp_rr --nolog -6 -c -H __GW6__%__IFNAME__ | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 200 150
|
|
|
|
tl TCP CRR latency over IPv6: guest to host
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
nsb tcp_crr --nolog -6
|
|
gout LAT tcp_crr --nolog -6 -c -H __GW6__%__IFNAME__ | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 500 400
|
|
|
|
tr TCP throughput over IPv4: guest to host
|
|
iperf3s ns 100${i}2 __THREADS__
|
|
|
|
guest ip link set dev __IFNAME__ mtu 256
|
|
iperf3 BW guest __GW__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 1M
|
|
bw __BW__ 0.2 0.3
|
|
guest ip link set dev __IFNAME__ mtu 576
|
|
iperf3 BW guest __GW__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 1M
|
|
bw __BW__ 0.5 0.8
|
|
guest ip link set dev __IFNAME__ mtu 1280
|
|
iperf3 BW guest __GW__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 4M
|
|
bw __BW__ 1.2 1.5
|
|
guest ip link set dev __IFNAME__ mtu 1500
|
|
iperf3 BW guest __GW__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 4M
|
|
bw __BW__ 1.6 1.8
|
|
guest ip link set dev __IFNAME__ mtu 9000
|
|
iperf3 BW guest __GW__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 8M
|
|
bw __BW__ 4.0 5.0
|
|
guest ip link set dev __IFNAME__ mtu 65520
|
|
iperf3 BW guest __GW__ 100${i}2 __THREADS__ __TIME__ __OPTS__ -w 16M
|
|
bw __BW__ 7.0 8.0
|
|
|
|
iperf3k ns
|
|
|
|
tl TCP RR latency over IPv4: guest to host
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
nsb tcp_rr --nolog -4
|
|
gout LAT tcp_rr --nolog -4 -c -H __GW__ | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 200 150
|
|
|
|
tl TCP CRR latency over IPv4: guest to host
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
nsb tcp_crr --nolog -4
|
|
gout LAT tcp_crr --nolog -4 -c -H __GW__ | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 500 400
|
|
|
|
tr TCP throughput over IPv6: host to guest
|
|
iperf3s guest 100${i}1 __THREADS__
|
|
|
|
bw -
|
|
bw -
|
|
bw -
|
|
bw -
|
|
bw -
|
|
iperf3 BW ns ::1 100${i}1 __THREADS__ __TIME__ __OPTS__
|
|
bw __BW__ 6.0 6.8
|
|
|
|
iperf3k guest
|
|
|
|
tl TCP RR latency over IPv6: host to guest
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
guestb tcp_rr --nolog -P 10001 -C 10011 -6
|
|
sleep 1
|
|
nsout LAT tcp_rr --nolog -P 10001 -C 10011 -6 -c -H ::1 | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 200 150
|
|
|
|
tl TCP CRR latency over IPv6: host to guest
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
guestb tcp_crr --nolog -P 10001 -C 10011 -6
|
|
sleep 1
|
|
nsout LAT tcp_crr --nolog -P 10001 -C 10011 -6 -c -H ::1 | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 500 350
|
|
|
|
|
|
tr TCP throughput over IPv4: host to guest
|
|
iperf3s guest 100${i}1 __THREADS__
|
|
|
|
bw -
|
|
bw -
|
|
bw -
|
|
bw -
|
|
bw -
|
|
iperf3 BW ns 127.0.0.1 100${i}1 __THREADS__ __TIME__ __OPTS__
|
|
bw __BW__ 6.0 6.8
|
|
|
|
iperf3k guest
|
|
|
|
tl TCP RR latency over IPv4: host to guest
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
guestb tcp_rr --nolog -P 10001 -C 10011 -4
|
|
sleep 1
|
|
nsout LAT tcp_rr --nolog -P 10001 -C 10011 -4 -c -H 127.0.0.1 | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 200 150
|
|
|
|
tl TCP CRR latency over IPv6: host to guest
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
lat -
|
|
guestb tcp_crr --nolog -P 10001 -C 10011 -4
|
|
sleep 1
|
|
nsout LAT tcp_crr --nolog -P 10001 -C 10011 -4 -c -H 127.0.0.1 | sed -n 's/^throughput=\(.*\)/\1/p'
|
|
lat __LAT__ 500 300
|
|
|
|
te
|