6.2. Ceph RBD performance report¶
Abstract: | This document includes Ceph RBD performance test results for 40 OSD nodes. Test cluster contain 40 OSD servers and forms 581TiB ceph cluster. |
---|
6.2.1. Environment description¶
Environment contains 3 types of servers:
- ceph-mon node
- ceph-osd node
- compute node
Role | Servers count | Type |
---|---|---|
compute | 1 | 1 |
ceph-mon | 3 | 1 |
ceph-osd | 40 | 2 |
6.2.1.1. Hardware configuration of each server¶
All servers have 2 types of configuration describing in table below
server | vendor,model | Dell PowerEdge R630 |
CPU | vendor,model | Intel,E5-2680 v3 |
processor_count | 2 | |
core_count | 12 | |
frequency_MHz | 2500 | |
RAM | vendor,model | Samsung, M393A2G40DB0-CPB |
amount_MB | 262144 | |
NETWORK | interface_name s | eno1, eno2 |
vendor,model | Intel,X710 Dual Port | |
bandwidth | 10G | |
interface_names | enp3s0f0, enp3s0f1 | |
vendor,model | Intel,X710 Dual Port | |
bandwidth | 10G | |
STORAGE | dev_name | /dev/sda |
vendor,model | raid1 - Dell, PERC H730P Mini
2 disks Intel S3610
|
|
SSD/HDD | SSD | |
size | 3,6TB |
server | vendor,model | Lenovo ThinkServer RD650 |
CPU | vendor,model | Intel,E5-2670 v3 |
processor_count | 2 | |
core_count | 12 | |
frequency_MHz | 2500 | |
RAM | vendor,model | Samsung, M393A2G40DB0-CPB |
amount_MB | 131916 | |
NETWORK | interface_names | enp3s0f0, enp3s0f1 |
vendor,model | Intel,X710 Dual Port | |
bandwidth | 10G | |
interface_names | ens2f0, ens2f1 | |
vendor,model | Intel,X710 Dual Port | |
bandwidth | 10G | |
STORAGE | vendor,model | 2 disks Intel S3610 |
SSD/HDD | SSD | |
size | 799GB | |
vendor,model | 10 disks 2T | |
SSD/HDD | HDD | |
size | 2TB |
6.2.1.2. Network configuration of each server¶
All servers have same network configuration:
6.2.1.3. Software configuration on servers with controller, compute and compute-osd roles¶
Ceph was deployed by Decapod tool. Cluster config for decapod:
ceph_config.yaml
Software | Version |
---|---|
Ceph | Jewel |
Ubuntu | Ubuntu 16.04 LTS |
You can find outputs of some commands and /etc folder in the following archives:
6.2.2. Testing process¶
- Run virtual machine on compute node with attached RBD disk.
- SSH into VM operation system
- Clone Wally repository.
- Create
ceph_raw.yaml
file in cloned repository - Run command python -m wally test ceph_rbd_2 ./ceph_raw.yaml
As a result we got the following HTML file: