Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Running Networking Performance Tests against K8s

License

NotificationsYou must be signed in to change notification settings

jtaleric/k8s-netperf

 
 

Repository files navigation

Running Networking Performance Tests against K8s

asciicast

Benchmark Tool and Tests

ToolTestStatusPass/Fail Context
netperfTCP_STREAMWorkingYes
netperfUDP_STREAMWorkingNo
netperfTCP_RRWorkingNo
netperfUDP_RRWorkingNo
netperfTCP_CRRWorkingNo
uperfTCP_STREAMWorkingYes
uperfUDP_STREAMWorkingNo
uperfTCP_RRWorkingNo
uperfUDP_RRWorkingNo
iperf3TCP_STREAMWorkingYes
iperf3UDP_STREAMWorkingNo

Setup

$ git clone http://github.com/cloud-bulldozer/k8s-netperf$cd k8s-netperf$ make build

Build Container Image

$ git clone http://github.com/cloud-bulldozer/k8s-netperf$cd k8s-netperf$ make container-build

Label nodes

k8s-netperf will make the best decision it can to schedule the client and server in your cluster. However,you can provide hints to ensure the client and server lands on specific nodes.

To do this, apply a label to the nodes you want the client and server running

$ oc label nodes node-name netperf=client$ oc label nodes node-name netperf=server

Running with Pods

Ensure yourkubeconfig is properly set to the cluster you would like to runk8s-netperf against.

also be sure to create anetperf namespace. (Not over-writable yet)

$ kubectl create ns netperf$ kubectl create sa -n netperf netperf

If you run with--all, you will need to allowhostNetwork for the netperf sa.

Example

$ oc adm policy add-scc-to-user hostnetwork -z netperf
$ kubectl create ns netperf$ kubectl create sa netperf -n netperf$ ./bin/amd64/k8s-netperf --helpA tool to run network performance testsin Kubernetes clusterUsage:  k8s-netperf [flags]Flags:      --config string         K8s netperf Configuration File (default"netperf.yml")      --netperf               Use netperf as load driver (default true)      --iperf                 Use iperf3 as load driver      --uperf                 Use uperf as load driver      --clean                 Clean-up resources created by k8s-netperf (default true)      --json                  Instead of human-readable output,return JSON to stdout      --local                 Run network performance tests with Server-Pods/Client-Pods on the same Node      --vm                    Launch Virtual Machines instead of podsfor client/servers      --vm-image string       Use specified VM image (default"kubevirt/fedora-cloud-container-disk-demo:latest")      --across                Place the client and server across availability zones      --all                   Run all tests scenarios - hostNet and podNetwork (if possible)      --debug                 Enable debug log      --udnl2                 Create and use a layer2 UDN as a primary network.      --udnl3                 Create and use a layer3 UDN as a primary network.      --udnPluginBinding string   UDN with VMs only - the binding method of the UDN interface,select'passt' or'l2bridge' (default"passt")      --serverIP string       External Server IP Addressfor the OCP client pod to communicate with.      --prom string           Prometheus URL      --uuid string           User provided UUID      --search string         OpenSearch URL,if you have auth, passin the format of https://user:pass@url:port      --index string          OpenSearch Index to save the results to, defaults to k8s-netperf      --metrics               Show all system metrics retrieved from prom      --tcp-tolerance float   Allowed %diff from hostNetwork to podNetwork, anything above tolerance will resultin k8s-netperf exiting 1. (default 10)      --version               k8s-netperf version      --csv                   Archive results, cluster and benchmark metricsin CSV files (default true)  -h, --helphelpfor k8s-netperf
  • --across will force the client to be across availability zones from the server
  • --json will reduce all output to just the JSON result, allowing users to feed the result tojq or other tools. Only output to the screen will be the result JSON or errors.
  • --clean=true will delete all the resources the project creates (deployments and services)
  • --serverIP accepts a string (IP Address). Example 44.243.95.221. k8s-netperf assumes this as server address and the client sends requests to this IP address.
  • --prom accepts a string (URL). Examplehttp://localhost:9090
    • When using--prom with a non-openshift cluster, it will be necessary to pass the prometheus URL.
  • --metrics will enable displaying prometheus captured metrics to stdout. By default they will be written to a csv file.
  • --iperf will enable the iperf3 load driver for any stream test (TCP_STREAM, UDP_STREAM). iperf3 doesn't have a RR or CRR test-type.
  • --uperf will enable the uperf load driver for any stream test (TCP_STREAM, UDP_STREAM). uperf doesn't have CRR test-type.

Note: With OpenShift, we attempt to discover the OpenShift route. If that route is not reachable, it might be required toport-forward the service and pass that via the--prom option.

Using External Server

This enables k8s-netperf to use the IP address provided via the--serverIP option as server address and the client sends requests to this IP address. This allows dataplane testing between ocp internal client pod and external server.

Note: User has to create a server with the provided IP address and run the intented k8s-netperf driver (i.e uperf, iperf or netperf). User has to enable respective ports on this server to allow the traffic from the client

Once the external server is ready to accept the traffic, users can orhestrate k8s-netperf by running

k8s-netperf --serverIP=44.243.95.221

Running with VMs

Running k8s-netperf against Virtual Machines (OpenShift CNV) requires

  • OpenShift CNV must be deployed and users should be able to define VMIs
  • SSH keys to be present in the home directory(~/.ssh/id_rsa.pub)
  • OpenShift Routes - k8s-netperf uses this to reach the VMs (k8s-netperf will create the route for the user, but we need Routes)

If the two above are in place, users can orhestrate k8s-netperf to launch VMs by running

k8s-netperf --vm

Using User Defined Network - UDN (only on OCP 4.18 and above)

To run k8s-netperf using a UDN primary network for the test instead of the default network of OVN-k:For a layer3 UDN:

$ k8s-netperf --udnl3

For a layer2 UDN:

$ k8s-netperf --udnl2

It works also with VMs

$ k8s-netperf --udnl2 --vm --udnPluginBinding=l2bridge

Warning! Support of k8s Services with UDN is not fully supported yet, you may faced inconsistent results when using a service in your tests.

Using a linux bridge interface

When using--bridge, a NetworkAttachmentDefinition defining a bridge interface is attached to the VMs and is used for the test. It requires the name of the bridge as it is defined in the NetworkNodeConfigurationPolicy, NMstate operator is required. For example:

apiVersion:nmstate.io/v1alpha1kind:NodeNetworkConfigurationPolicymetadata:name:br0-eth1spec:desiredState:interfaces:      -name:br0description:Linux bridge with eno2 as a porttype:linux-bridgestate:upipv4:dhcp:trueenabled:truebridge:options:stp:enabled:falseport:            -name:eno2

Then you can launch a test using the bridge interface:

./bin/amd64/k8s-netperf --vm --bridge br0

By default, it will read thebridgeNetwork.json file from the git repository. If the default IP addresses (10.10.10.12/24 and 10.10.10.14/24) are not available for your setup, it is possible to change it by passing a JSON file as a parameter with--bridgeNetwork, like follow:

k8s-netperf --vm --bridge br0 --bridgeNetwork /path/to/my/bridgeConfig.json

Config file

Config File v2

The v2 config file will be executed in the order the tests are presented in the config file.

tests:  -TCPStream:# Place-holder of a test nameparallelism:1# Number of concurrent netperf processes to run.profile:"TCP_STREAM"# Netperf profile to execute. This can be [TCP,UDP]_STREAM, [TCP,UDP]_RR, TCP_CRRduration:3# How long to run the testsamples:1# Iterations to run specified testmessagesize:1024# Size of the data-gramburst:1# Number of transactions inflight at one time. By default, netperf does one transaction at a time. This is netperf's TCP_RR specific option.service:false# If we should test with the server pod behind a service

Config File v1

The v1 config file will be executed in random order. If you need the list of tests to be ran in a specific order, use the v2 config file format.netperf.yml contains a default set of tests.

Description of each field in the YAML.

TCPStream:# Place-holder of a test nameparallelism:1# Number of concurrent netperf processes to run.profile:"TCP_STREAM"# Netperf profile to execute. This can be [TCP,UDP]_STREAM, [TCP,UDP]_RR, TCP_CRRduration:3# How long to run the testsamples:1# Iterations to run specified testmessagesize:1024# Size of the data-gramburst:1# Number of transactions inflight at one time. By default, netperf does one transaction at a time. This is netperf's TCP_RR specific option.service:false# If we should test with the server pod behind a service

Parallelism

In most cases setting parallelism greater than 1 is OK, when usingservice: true, multiple threads (or processes in netperf) connect to the same service.

Pass / Fail

k8s-netperf has a cli option for--tcp-tolerance which defaults to 10%.

In order to havek8s-netperf determine pass/fail the user must pass the--all flag.k8s-netperf must be able to run with hostNetwork and podNetwork across nodes.

$ ./k8s-netperf --tcp-tolerance 1+-------------------+---------+------------+-------------+--------------+---------+-----------------+-------+-----------+----------+---------+--------------------+|    RESULT TYPE| DRIVER|  SCENARIO| PARALLELISM| HOST NETWORK| SERVICE| EXTERNAL SERVER| MESSAGE SIZE| BURST| SAME NODE| DURATION| SAMPLES|     AVG VALUE|+-------------------+---------+------------+-------------+--------------+---------+-----------------+-------+-----------+----------+---------+--------------------+| 📊 Stream Results| netperf| TCP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 2661.006667 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 2483.078229 (Mb/s)|| 📊 Stream Results| uperf| TCP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 2581.705097 (Mb/s)|| 📊 Stream Results| netperf| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 2702.230000 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 2523.434069 (Mb/s)|| 📊 Stream Results| uperf| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 2567.665412 (Mb/s)|| 📊 Stream Results| netperf| TCP_STREAM| 1|true|false|false| 8192| 0|false| 10| 3| 2697.276667 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|true|false|false| 8192| 0|false| 10| 3| 2542.793728 (Mb/s)|| 📊 Stream Results| uperf| TCP_STREAM| 1|true|false|false| 8192| 0|false| 10| 3| 2571.881579 (Mb/s)|| 📊 Stream Results| netperf| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 2707.076667 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 2604.067072 (Mb/s)|| 📊 Stream Results| uperf| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 2687.276667 (Mb/s)|| 📊 Stream Results| netperf| UDP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 1143.926667 (Mb/s)|| 📊 Stream Results| iperf3| UDP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 1202.428288 (Mb/s)|| 📊 Stream Results| uperf| UDP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 1242.059988 (Mb/s)|| 📊 Stream Results| netperf| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 1145.066667 (Mb/s)|| 📊 Stream Results| iperf3| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 1239.580672 (Mb/s)|| 📊 Stream Results| uperf| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 1261.840000 (Mb/s)|+-------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+--------------------++---------------+---------+----------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+---------------------+|  RESULT TYPE| DRIVER| SCENARIO| PARALLELISM| HOST NETWORK| SERVICE| EXTERNAL SERVER| MESSAGE SIZE| BURST| SAME NODE| DURATION| SAMPLES|      AVG VALUE|+---------------+---------+----------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+---------------------+| 📊 Rr Results| netperf| TCP_CRR| 1|true|true|false| 1024| 0|false| 10| 3| 2370.196667 (OP/s)|| 📊 Rr Results| netperf| TCP_CRR| 1|false|true|false| 1024| 0|false| 10| 3| 3046.126667 (OP/s)|| 📊 Rr Results| netperf| TCP_RR| 1|true|false|false| 1024| 2|false| 10| 3| 16849.056667 (OP/s)|| 📊 Rr Results| netperf| TCP_RR| 1|false|false|false| 1024| 2|false| 10| 3| 17101.856667 (OP/s)|| 📊 Rr Results| netperf| TCP_CRR| 1|true|false|false| 1024| 0|false| 10| 3| 3166.136667 (OP/s)|| 📊 Rr Results| netperf| TCP_CRR| 1|false|false|false| 1024| 0|false| 10| 3| 1787.530000 (OP/s)|+---------------+---------+----------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+---------------------++---------------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+-----------------------------+|        RESULT TYPE| DRIVER|  SCENARIO| PARALLELISM| HOST NETWORK| SERVICE| EXTERNAL SERVER| MESSAGE SIZE| BURST| SAME NODE| DURATION| SAMPLES|   99%TILE VALUE|+---------------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-------------------+| 📊 Stream Latency Results| netperf| TCP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 71.333333 (usec)|| 📊 Stream Latency Results| netperf| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 2.333333 (usec)|| 📊 Stream Latency Results| netperf| TCP_STREAM| 1|true|false|false| 8192| 0|false| 10| 3| 276.000000 (usec)|| 📊 Stream Latency Results| netperf| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 124.333333 (usec)|| 📊 Stream Latency Results| netperf| UDP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 14.666667 (usec)|| 📊 Stream Latency Results| netperf| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 14.666667 (usec)|+---------------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-------------------++-----------------------+---------+----------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-------------------+|      RESULT TYPE| DRIVER| SCENARIO| PARALLELISM| HOST NETWORK| SERVICE| EXTERNAL SERVER| MESSAGE SIZE| BURST| SAME NODE| DURATION| SAMPLES|   99%TILE VALUE|+-----------------------+---------+----------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-------------------+| 📊 Rr Latency Results| netperf| TCP_CRR| 1|true|true|false| 1024| 0|false| 10| 3| 817.333333 (usec)|| 📊 Rr Latency Results| netperf| TCP_CRR| 1|false|true|false| 1024| 0|false| 10| 3| 647.666667 (usec)|| 📊 Rr Latency Results| netperf| TCP_RR| 1|true|false|false| 1024| 2|false| 10| 3| 125.333333 (usec)|| 📊 Rr Latency Results| netperf| TCP_RR| 1|false|false|false| 1024| 2|false| 10| 3| 119.666667 (usec)|| 📊 Rr Latency Results| netperf| TCP_CRR| 1|true|false|false| 1024| 0|false| 10| 3| 621.000000 (usec)|| 📊 Rr Latency Results| netperf| TCP_CRR| 1|false|false|false| 1024| 0|false| 10| 3| 539.666667 (usec)|+-----------------------+---------+----------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-------------------+😥 TCP Stream percent difference when comparing hostNetwork to podNetwork is greater than 1.0 percent (2.7 percent)$echo$?1
ToolTestpass/fail
netperfTCP_STREAMworking (default:10%)

Indexing to OpenSearch

k8s-netperf can store results in OpenSearch, if the user provides the OpenSearch URL.

rhinoin~ $ ./k8s-netperf --config test.yml --search https://admin:pass@my-es:443...<trimmed output>INFO[2023-03-02 16:38:48] Connected to: [https://admin:pass@my-es:443] INFO[2023-03-02 16:38:48] Attempting to index 2 documents

Document format can be seen inpkg/archive/archive.go

Output

k8s-netperf will provide updates to stdout of the operations it is running, such as creating the server/client deployments and the execution of the workload in the container.

Same node refers to how the pods were deployed. If the cluster has > 2 nodes with nodes which haveworker= there will be a cross-node throughput test.

+-------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+--------------------+|    RESULT TYPE| DRIVER|  SCENARIO| PARALLELISM| HOST NETWORK| SERVICE| EXTERNAL SERVER| MESSAGE SIZE| BURST| SAME NODE| DURATION| SAMPLES|     AVG VALUE|+-------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+--------------------+| 📊 Stream Results| netperf| TCP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 2661.006667 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 2483.078229 (Mb/s)|| 📊 Stream Results| netperf| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 2702.230000 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 2523.434069 (Mb/s)|| 📊 Stream Results| netperf| TCP_STREAM| 1|true|false|false| 8192| 0|false| 10| 3| 2697.276667 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|true|false|false| 8192| 0|false| 10| 3| 2542.793728 (Mb/s)|| 📊 Stream Results| netperf| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 2707.076667 (Mb/s)|| 📊 Stream Results| iperf3| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 2604.067072 (Mb/s)|| 📊 Stream Results| netperf| UDP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 1143.926667 (Mb/s)|| 📊 Stream Results| iperf3| UDP_STREAM| 1|true|false|false| 1024| 0|false| 10| 3| 1202.428288 (Mb/s)|| 📊 Stream Results| netperf| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 1145.066667 (Mb/s)|| 📊 Stream Results| iperf3| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 1239.580672 (Mb/s)|+-------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+--------------------+

Loss/Retransmissions

k8s-netperf will report TCP Retransmissions and UDP Loss for both workload drivers (netperf and iperf).

+---------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-----------+|        TYPE| DRIVER|  SCENARIO| PARALLELISM| HOST NETWORK| SERVICE| EXTERNAL SERVER| MESSAGE SIZE| BURST| SAME NODE| DURATION| SAMPLES| AVG VALUE|+---------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-----------+| TCP Retransmissions| netperf| TCP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 54.666667|| TCP Retransmissions| netperf| TCP_STREAM| 1|false|false|false| 8192| 0|false| 10| 3| 15.000000|| UDP Loss Percent| netperf| UDP_STREAM| 1|false|false|false| 1024| 0|false| 10| 3| 0.067031|+---------------------+---------+------------+-------------+--------------+---------+--------------+-------+-----------+----------+---------+-----------+

Output to CSV

k8s-netperf will write a csv file, after it has completed the desired performance tests.

Example output

Driver,Profile,Same node,Host Network,Service,External Server,Duration,Parallelism,# of Samples,Message Size,Confidence metric - low,Confidence metric - high,Avg Throughput,Throughput Metric,99%tile Observed Latency,Latency Metricnetperf,TCP_STREAM,false,false,false,false,10,1,3,1024,861.9391413991156,885.2741919342178,873.606667,Mb/s,3.3333333333333335,usecnetperf,TCP_STREAM,false,false,false,false,10,1,3,8192,178.12442996547009,1310.3422367011967,744.233333,Mb/s,2394.6666666666665,usecnetperf,UDP_STREAM,false,false,false,false,10,1,3,1024,584.3478157889886,993.4588508776783,788.903333,Mb/s,23,usecnetperf,TCP_CRR,false,false,false,false,10,1,3,1024,1889.3183973002176,2558.074936033115,2223.696667,OP/s,4682.666666666667,usecnetperf,TCP_CRR,false,false,true,false,10,1,3,1024,1169.206855676418,2954.3464776569153,2061.776667,OP/s,4679.333333333333,usecnetperf,TCP_RR,false,false,false,false,10,1,3,1024,6582.5359452538705,12085.437388079461,9333.986667,OP/s,451.3333333333333,usec

About

Running Networking Performance Tests against K8s

Resources

License

Code of conduct

Contributing

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Go96.2%
  • Shell1.6%
  • Makefile1.4%
  • Dockerfile0.8%

[8]ページ先頭

©2009-2025 Movatter.jp