6.18.3. OpenStack Neutron Resource Density Test Report

This report is generated for OpenStack Neutron Resource Density Test Plan.

6.18.3.1. Environment description

6.18.3.1.1. Cluster description

  • 3 controllers

  • 2 compute nodes

6.18.3.1.2. Software versions

OpenStack/System:

Fuel/MOS 9.0, Ubuntu 14.04, Linux kernel 3.13, OVS 2.4.1

Networking

Neutron ML2 + OVS plugin, DVR, L2pop, MTU 1500

6.18.3.1.3. Hardware configuration of each server

Description of servers hardware

Compute Vendor:

HP ProLiant DL380 Gen9,

CPU

2 x Intel(R) Xeon(R) CPU E5-2680 v3 @2.50GHz (48 cores)

RAM:

256 Gb

NIC

2 x Intel Corporation Ethernet 10G 2P X710

6.18.3.2. Report

6.18.3.2.1. Preparation

Measurements are performed between 2 instances deployed in L3 east-west topology. To prepare:

  1. Create new Neutron router. Plug into external network.

  2. Create 2 Neutron networks, e.g. east and west. Plug them into the router.

  3. Boot 2 instances (Ubuntu Cloud Image can be used). Each instance should be in its own network and on its own compute node. Public key may be required to login into instance (in case of using Ubuntu).

  4. Assign floating IP to one of instances.

The traffic is sent between 2 instances, from the one with floating IP to the other. Phisically packets will go from one compute node to the other directly.

6.18.3.2.2. Measurements

6.18.3.2.2.1. Baseline

We start with measuring baseline numbers. During this test only 2 instances exist.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

1.8 [0.6]

7445 [850]

132 [6.5]

6.18.3.2.2.2. 10 instances per compute in separate network

There are 20 micro instances (10 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Instances are plugged into separate network.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.0 [0.7]

6500 [807]

106 [4.5]

6.18.3.2.2.3. 30 instances per compute in separate network

There are 60 micro instances (30 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Instances are plugged into separate network.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.2 [0.8]

6650 [600]

88 [3.0]

6.18.3.2.2.4. 10 instances per compute in tester network

There are 20 micro instances (10 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Instances are plugged into east network.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.7 [1.0]

4930 [499]

101 [3.6]

6.18.3.2.2.5. 30 instances per compute in tester network

There are 60 micro instances (30 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Instances are plugged into east network.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

5.4 [2.1]

2470 [130]

88 [2.3]

6.18.3.2.2.6. 10 instances and 10 routers per compute

There are 20 micro instances (10 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Each instance is plugged into its own network, which in turn is plugged into its own router.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.3 [0.8]

5690 [630]

107 [5.0]

6.18.3.2.2.7. 30 instances and 30 routers per compute

There are 60 micro instances (30 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Each instance is plugged into its own network, which in turn is plugged into its own router.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.2 [0.8]

5890 [707]

85 [1.9]

6.18.3.2.2.8. 30 instances and 30 routers per compute, QEMU paused

There are 60 micro instances (30 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Each instance is plugged into its own network, which in turn is plugged into its own router. QEMU processes are paused to reduce noise load on CPU.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.0 [0.7]

6640 [623]

103 [4.6]

6.18.3.2.2.9. 30 instances, 30 routers and 500 SG rules per instance, QEMU paused

There are 60 micro instances (30 per compute node). Each instance is Cirros-based with 1 VCPU and 64MB RAM. Each instance is plugged into its own network, which in turn is plugged into its own router. QEMU processes are paused to reduce noise load on CPU. All instances have security group which contains 500 rules (this results in ~17k iptables rules)

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

2.0 [0.7]

6658 [576]

97 [1.0]

6.18.3.2.2.10. 1000 security group rules

In this case we assign security group containing 1000 rules to test instances. There is no measurable difference from the baseline data.

Latency, ms [σ]

TCP throughput, Mbits/s [σ]

UDP throughput, kpps [σ]

1.8 [0.6]

7039 [653]

130 [2.0]

6.18.3.2.2.11. Conclusion

It appears that two things have most of impact:
  • neighbours, which produce slight CPU load

  • OVS flows: the more flows, the less performance. Adding 30 instances to compute node results in 180 new flows.

../../../_images/chart_latency.png

The worst case for latency is when instances are co-located in the same network as the tester. While presence of security group rules have no measurable effect at all.

../../../_images/chart_tcp.png

The worst case again when instances are co-located in the same network. CPU noise also affects - the throughput reduces on ~10%.

../../../_images/chart_udp.png

For UDP metric the most impact is introduced by CPU noise and presence of OVS rules. With security group rules the numbers are the same as in the baseline.