Test Results for fuel-os-nosdn-nofeature-ha


Overview of test results

See Grafana for viewing test result metrics for each respective test case. It is possible to chose which specific scenarios to look at, and then to zoom in on the details of each run test scenario as well.

All of the test case results below are based on 5 consecutive scenario test runs, each run on the Ericsson POD2 between February 13 and 18 in 2016. The best would be to have more runs to draw better conclusions from, but these are the only runs available at the time of OPNFV R2 release.


The round-trip-time (RTT) between 2 VMs on different blades is measured using ping. The measurements are on average varying between 0.5 and 1.1 ms with a first 2 - 2.5 ms RTT spike in the beginning of each run (This could be because of normal ARP handling). The 2 last runs are very similar in their results. But, to be able to draw any further conclusions more runs should be made. There is one measurement taken on February 16 that does not have the first RTT spike, and less variations to the RTT. The reason for this is unknown. There is a discussion on another test measurement made Feb. 16 in TC037. SLA set to 10 ms. The SLA value is used as a reference, it has not been defined by OPNFV.


The IO read bandwidth look similar between different test runs, with an average at approx. 160-170 MB/s. Within each run the results vary much, minimum 2 MB/s and maximum 630 MB/s on the totality. Most runs have a minimum of 3 MB/s (one run at 2 MB/s). The maximum BW varies much more in absolute numbers, between 566 and 630 MB/s. SLA set to 400 MB/s. The SLA value is used as a reference, it has not been defined by OPNFV.


The measurements for memory latency are consistent among test runs and results in approx. 1.2 ns. The variations between runs are similar, between 1.215 and 1.219 ns. One exception is February 16, where the varation is greater, between 1.22 and 1.28 ns. SLA set to 30 ns. The SLA value is used as a reference, it has not been defined by OPNFV.


For this scenario no results are available to report on. Probable reason is an integer/floating point issue regarding how InfluxDB is populated with result data from the test runs.


The average measurements for memory bandwidth are consistent among most of the different test runs at 17.2 - 17.3 GB/s. The very first test run averages at 17.7 GB/s. Within each run the results vary, with a minimal BW of 15.4 GB/s and maximum of 18.2 GB/s of the totality. SLA set to 15 GB/s. The SLA value is used as a reference, it has not been defined by OPNFV.


The Unixbench processor single and parallel speed scores show similar results at approx. 3200. The runs vary between scores 3160 and 3240. No SLA set.


The amount of packets per second (PPS) and round trip times (RTT) between 2 VMs on different blades are measured when increasing the amount of UDP flows sent between the VMs using pktgen as packet generator tool.

Round trip times and packet throughput between VMs are typically affected by the amount of flows set up and result in higher RTT and less PPS throughput.

When running with less than 10000 flows the results are flat and consistent. RTT is then approx. 30 ms and the number of PPS remains flat at approx. 250000 PPS. Beyond approx. 10000 flows and up to 1000000 (one million) there is an even drop in RTT and PPS performance, eventually ending up at approx. 150-250 ms and 40000 PPS respectively.

There is one measurement made February 16 that has slightly worse results compared to the other 4 measurements. The reason for this is unknown. For instance anyone being logged onto the POD can be of relevance for such a disturbance.

Detailed test results

The scenario was run on Ericsson POD2 with: Fuel 8.0 OpenStack Liberty OVS 2.3.1

No SDN controller installed

Rationale for decisions


Tests were successfully executed and metrics collects (apart from TC011). No SLA was verified. To be decided on in next release of OPNFV.

Conclusions and recommendations

The pktgen test configuration has a relatively large base effect on RTT in TC037 compared to TC002, where there is no background load at all (30 ms compared to 1 ms or less, which is more than a 3000 percentage different in RTT results). The larger amounts of flows in TC037 generate worse RTT results, in the magnitude of several hundreds of milliseconds. It would be interesting to also make and compare all these measurements to completely (optimized) bare metal machines running native Linux with all other relevant tools available, e.g. lmbench, pktgen etc.