5.1. 1000 Compute nodes resource consumption/scalability testing¶
- status
ready
- version
1.0
- Abstract
This document describes a test plan for measuring OpenStack services resources consumption along with scalability potential. It also provides a results which could be used to find bottlenecks and/or potential pain points for scaling standalone OpenStack services and OpenStack cloud itself.
5.1.1. Test Plan¶
Most of current OpenStack users wonder how it will behave on scale with a lot of compute nodes. This is a valid concern because OpenStack have a lot of services whose have different load and resources consumptions patterns. Most of the cloud operations are related to the two things: workloads placement and simple control/data plane management for them. So the main idea of this test plan is to create simple workloads (10-30k of VMs) and observe how core services working with them and what is resources consumption during active workloads placement and some time after that.
5.1.1.1. Test Environment¶
Test assumes that each and every service will be monitored separately for resources consuption using known techniques like atop/nagios/containerization and any other toolkits/solutions which will allow to:
Measure CPU/RAM consumption of process/set of processes.
Separate services and provide them as much as possible resources available to fulfill their needs.
- List of mandatory services for OpenStack testing:
nova-api nova-scheduler nova-conductor nova-compute glance-api glance-registry neutron-server keystone-all
- List of replaceable but still mandatory services:
neutron-dhcp-agent neutron-ovs-agent rabbitmq libvirtd mysqld openvswitch-vswitch
- List of optional service which may be omitted with performance decrease:
memcached
- List of optional service which may be omitted:
horizon
Rally fits here as a pretty stable and reliable load runner. Monitoring could be done by any suitable software which will be able to provide a results in a form which allow to build graphs/visualize resources consumption to analyze them or do the analysis automatically.
5.1.1.1.1. Preparation¶
Common preparation steps
To begin testing environment should have all the OpenStack services up and running. Of course they should be configured accordingly to the recommended settings from release and/or for your specific environment or use case. To have real world RPS/TPS/etc metrics all the services (including compute nodes) should be on the separate physical servers but again it depends on setup and requirements. For simplicity and testing only control plane the Fake compute driver could be used.
5.1.1.1.2. Environment description¶
The environment description includes hardware specification of servers, network parameters, operation system and OpenStack deployment characteristics.
5.1.1.1.2.1. Hardware¶
This section contains list of all types of hardware nodes.
Parameter |
Value |
Comments |
model |
e.g. Supermicro X9SRD-F |
|
CPU |
e.g. 6 x Intel(R) Xeon(R) CPU E5-2620 v2 @ 2.10GHz |
5.1.1.1.2.2. Network¶
This section contains list of interfaces and network parameters. For complicated cases this section may include topology diagram and switch parameters.
Parameter |
Value |
Comments |
card model |
e.g. Intel |
|
driver |
e.g. ixgbe |
|
speed |
e.g. 10G or 1G |
5.1.1.1.2.3. Software¶
This section describes installed software.
Parameter |
Value |
Comments |
OS |
e.g. Ubuntu 14.04.3 |
|
DB |
e.g. MySQL 5.6 |
|
MQ broker |
e.g. RabbitMQ v3.4.25 |
|
OpenStack release |
e.g. Liberty |
5.1.1.1.2.4. Configuration¶
This section describes configuration of OpenStack and core services
Parameter |
File |
Keystone |
./results/keystone.conf |
Nova-api |
./results/nova-api.conf |
… |
|
5.1.1.2. Test Case 1: Resources consumption under severe load¶
5.1.1.2.1. Description¶
This test should spawn a number of instances in n parallel threads and along with that record all CPU/RAM metrics from all the OpenStack and core services like MQ brokers and DB server. As test itself is pretty long there is no need in very high test resolution. 1 measure per 5 seconds should be more than enough.
Rally scenario that creates load of 50 parallel threads spawning VMs and calling for VMs list can be found in test plan folder and can be used for testing purposes. It could be modified to fit specific deployment needs.
5.1.1.2.2. Parameters¶
Parameter name |
Value |
---|---|
OpenStack release |
Liberty, Mitaka |
Compute nodes amount |
50,100,200,500,1000,2000,5000,10000 |
Services configurations |
Configuration for each OpenStack and core service |
5.1.1.2.3. List of performance metrics¶
Test case result is presented as a weighted tree structure with operations as nodes and time spent on them as node weights for every control plane operation under the test. This information is automatically gathered in Ceilometer and can be gracefully transformed to the human-friendly report via OSprofiler.
Priority |
Value |
Measurement Units |
Description |
---|---|---|---|
1 |
CPU load |
Mhz |
CPU load for each OpenStack service |
2 |
RAM consumption |
Gb |
RAM consumption for each OpenStack service |
3 |
Instances amnt |
Amount |
Max number of instances spawned |
4 |
Operation time |
milliseconds |
Time spent for every instance spawn |
5.1.2. Reports¶
- Test plan execution reports: