SlideShare a Scribd company logo
Red Hat Ceph
Performance & Sizing
Guide
Jose De la Rosa
September 7, 2016
2
Agenda
1. Ceph Overview
a. Architecture
b. Approach to storing data
2. Test methodology
a. What was tested
b. How it was tested & measured
3. Results
a. Key findings
b. Recommendations
3
Ceph Overview
4
Ceph Overview
1. Open source software defined storage: logical storage services and
capabilities are abstracted from the underlying physical storage
systems.
2. Provides object, block and file system storage. All data is stored as
objects.
3. Massively scalable to thousands of storage nodes.
4. Self-healing with no single point of failure: If a node fails, it is
automatically detected and data rebalances to ensure availability.
5
Client access
Ross Turk, Red Hat - https://raw.githubusercontent.com/ceph/ceph/master/doc/images/stack.png
6
Architecture
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSD
Disk
OSDObject
Storage
Devices
PG
PG
PG
PG
CRUSH algorithm
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
PG
Pools &
Placement
Groups
Obj Obj
Obj
Obj
Obj
Obj
Obj
Obj
RADOS
Object (RGW) Block (librbd) CephFS
Client
Interface
Layer
Objects
Obj
Obj
Obj
Monitors
Monitor
Monitor
Monitor
Metadata
Servers
MDS
MDS
MDS
7
Storage Protection Method
Disk Disk Disk
OSD OSD OSD
3x Replication
Data copied 3 times
Actual disk usage = 33%
Disk Disk Disk Disk
OSD OSD OSD OSD
Erasure Coded (3+2)
Data split into 3 + 2 disks used for parity
Actual disk usage = 60%
Disk
OSD
8
Reading and Writing Data
http://www.slideshare.net/LarryCover/ceph-open-source-storage-software-optimizations-on-intel-architecture-for-cloud-workloads
9
Test Setup &
Methodology
10
Server configuration
Testbed Details
Ceph tier Storage Nodes (5) Monitors (3) Clients (10)
Platform Dell PowerEdge R730xd Dell PowerEdge R630 Dell PowerEdge R220
CPU 2x Intel Xeon E5-2630 v3 2.4GHz 2x Intel Xeon E5-2650 v3 2.3 GHz 1x Intel Celeron G1820 2.7 GHz
Memory 4x 16 GB 1866 MHz DDR4 8x 16 GB 2133MHz DDR4 4x 4 GB 1600 MHz DDR3
Network 1x Intel X520/2P I350 LOM 1x Intel X520/2P I350 LOM 1x Intel X520/2P I350
Storage PERC H730 Mini / 1 GB Cache
Up to 16x: SEAGATE 4 TB SAS
(ST4000NM0005)
Up to 3x: Intel DC S3700 SSD 200
GB SATA (SSDSC2BA20)
1x Intel DC P3700 SSD 800 GB
NVMe
PERC H730 Mini / 1 GB Cache
6x SEAGATE 500 GB SAS
(ST9500620S)
1x Toshiba 50 GB SATA (DT01ACA0)
11
Network Topology
12
Configuration Guidance
1. General rules of thumb
• 1 Core-GHz per OSD
• SATA/SAS SSD to HDD Ratio: 1:4 - 1:5
• NVME SSD to HDD Ratio: 1:17-1:18
• 16GB RAM Baseline + 2-3GB per OSD
2. More details at https://www.redhat.com/en/resources/red-hat-
ceph-storage-hardware-configuration-guide
13
Storage node configurations tested
OSD to Journal
Ratio [drives]
12+3 16+0 16+1
OSD node
configuration
12+3 16+0 16+1
HDDs 12 16 16
HDD RAID mode Single-disk RAID0 Single-disk RAID0 Single-disk RAID0 / HBA mode
SATA SSDs 3 0 0
SSD RAID mode JBOD JBOD JBOD
NVMe SSDs 0 0 1
Network 1x 10 GbE Front-End
1x 10 GbE Back-End
1x 10 GbE Front-End
1x 10 GbE Back-End
1x 10 GbE Front-End
1x 10 GbE Back-End
14
Benchmarking with CBT
1. For benchmark automation, the open source utility Ceph
Benchmarking Tool (CBT) was used.
2. It supports different drivers for examining different layers of the
storage stack:
• radosbench - uses librados API (used in this study)
• librbdfio – test block storage without KVM/QEMU instances
• kvmrbdfio – test block volumes attached to KVM/QEMU instances
3. Available at https://github.com/ceph/cbt
15
Factors that influence performance
1. Device used for journaling (SSD vs. HDD)
2. RAID0 vs. pass-through (HBA) mode
3. Number of clients (single stream vs. parallel access)
4. Data-protection mode (3-way replication vs. erasure coding)
16
Test Matrix
Server configuration
PowerEdge R730xd
12+3, 3xRep
PowerEdge R730xd
16+0, EC3+2
PowerEdge R730xd
16r+1, 3xRep
PowerEdge R730xd
16+1, EC 3+2
PowerEdge R730xd
16j+1, 3xRep
OS disk 2x 500 GB 2.5" 2x 500 GB 2.5" 2x 500 GB 2.5" 2x 500 GB 2.5" 2x 500 GB 2.5"
Data disk type
HDD 7.2K SAS 12Gbps,
4TB
HDD 7.2K SAS 12Gbps,
4TB
HDD 7.2K SAS 12Gbps,
4TB
HDD 7.2K SAS 12Gbps,
4TB
HDD 7.2K SAS 12Gbps,
4TB
HDD quantity 12 16 16 16 16
Number of Ceph write
journal devices
3 0 1 1 1
Ceph write journal device
type
Intel SATA SSD S3710
(6Gb/s)
n/a
Intel P3700 PCIe NVMe
HHHL AIC
Intel P3700 PCIe NVMe
HHHL AIC
Intel P3700 PCIe NVMe
HHHL AIC
Ceph write journal device
size (GB)
200 0 800 800 800
Controller model
PERC H730,
1 GB Cache
PERC H730,
1 GB Cache
PERC H730,
1 GB Cache
PERC H730,
1 GB Cache
PERC H730,
1 GB Cache
PERC Controller
configuration for HDDs
RAID RAID RAID RAID
JBOD
(PERC pass-through
mode)
Raw capacity for Ceph
OSDs (TB)
48 64 64 64 64
17
Benchmark Test
Results
18
Throughput / server
0 200 400 600 800 1000 1200 1400
R730xd 16r+1, 3xRep
R730xd 16j+1, 3xRep
R730XD 16+1, EC3+2
R730xd 16+0, EC3+2
R730xd 12+3, 3xRep
MBps per Server (4MB seq IO)
Performance Comparison
Throughput per Server
Writes
Reads
19
Overall Solution Price/Performance
R730xd 16+0, EC3+2
R730xd 16+0, EC8+3
R730xd 16r+1, 3xRep
R730xd 16j+1, 3xRep
R730XD 16+1, EC3+2
R730xd 16+1, EC8+3
R730xd 12+3, 3xRep
Total Cluster Server+SW Cost / Cluster MBps
Solution Price/Performance Comparison
500TB Usable Cluster
(less $ per MBps is better)
Write
Read
$0
20
Overall Solution Price Capacity
R730xd 16+0, EC3+2
R730xd 16+0, EC8+3
R730xd 16r+1, 3xRep
R730xd 16j+1, 3xRep
R730XD 16+1, EC3+2
R730xd 16+1, EC8+3
R730xd 12+3, 3xRep
Total Cluster Server+SW Cost / Cluster GB
Solution Price/Capacity Comparison
(less $ per GB is better)
$0
21
Replication vs. Erasure Coding
0 200 400 600 800 1000 1200 1400
R730xd 16r+1, 3xRep
R730xd 16j+1, 3xRep
R730XD 16+1, EC3+2
R730xd 16+1, EC8+3
MBps per Server (4MB seq IO)
Performance Comparison
Replication vs. Erasure-coding
Writes
Reads
22
JBOD vs. RAID0
1153
1147
393
342
0 200 400 600 800 1000 1200 1400
R730XD 16 RAID+1, 3xRep
R730XD 16 JBOD+1, 3xRep
MBps per Server (4MB seq IO)
Performance Comparison
JBOD vs. RAID0 Config
Writes
Reads
23
Performance conclusions
1. Replication mode yielded better performance for read operations
and the erasure-coded mode proved better for write operations.
2. The PowerEdge R730xd 16+1 3x replication configuration yielded
optimal price for read-write throughput-oriented workloads.
3. The PowerEdge R730xd 12+3 3x replication configuration yielded
optimal price for read-only throughput-oriented workloads.
4. The PowerEdge R730xd 16+1 erasure-coded configuration proved
to be the choice for write-heavy operations.
5. When used with Ceph Storage, Dell & Red Hat recommend the
usage of single-drive RAID0 mode on PowerEdge R730xd with PERC
H730.
24
Sizing Recommendations
Storage Capacity Extra Small Small Medium
Cluster Capacity 100 TB+ 500 TB+ 1 PB+
Throughput-Optimized >4x R730xd (8U) >8x R730xd (16U) NA
1x server/2U chassis 1x server/2U chassis
16x 6 TB HDD 16x 6 TB HDD
1x 800 GB NVMe SSD 1x 800 GB NVMe SSD
2x 10 GbE 2x 10 GbE
3x Replication 3x Replication
Cost/Capacity-Optimized NA NA >15x R730xd (30U)
1x server/2U chassis
16x 8 TB HDD
1x HHHL AIC SSD
2x 10 GbE
8:3 Erasure-coding
25
Observations
1. Obey SSD to HDD ratio
2. Hardware matters, look at RAID controllers if you use HDDs
3. Don’t use RAID controllers on SSDs
4. SSD sequential write bandwidth becomes a bottleneck
5. Random workloads should go on Flash-only
6. 10GbE Bonding not necessary with <=16 drives
26
Recommended Reading
Dell PowerEdge R730xd Red Hat Ceph Storage Performance and
Sizing Guide
http://en.community.dell.com/techcenter/cloud/m/dell_cloud_resources/20442913
Enterprise Ceph: Every Way, Your Way
https://www.redhat.com/files/summit/session-assets/2016/SS88828-enterprise-
ceph_every-way-your-way.pdf
Ad

More Related Content

What's hot (20)

BlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InBlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year In
Sage Weil
 
ceph optimization on ssd ilsoo byun-short
ceph optimization on ssd ilsoo byun-shortceph optimization on ssd ilsoo byun-short
ceph optimization on ssd ilsoo byun-short
NAVER D2
 
2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph
Ceph Community
 
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake SolutionCeph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Karan Singh
 
Ceph Block Devices: A Deep Dive
Ceph Block Devices:  A Deep DiveCeph Block Devices:  A Deep Dive
Ceph Block Devices: A Deep Dive
Red_Hat_Storage
 
Ceph
CephCeph
Ceph
Hien Nguyen Van
 
Ceph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing GuideCeph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing Guide
Karan Singh
 
Ceph scale testing with 10 Billion Objects
Ceph scale testing with 10 Billion ObjectsCeph scale testing with 10 Billion Objects
Ceph scale testing with 10 Billion Objects
Karan Singh
 
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA ArchitectureCeph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Danielle Womboldt
 
Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)
Sage Weil
 
Nick Fisk - low latency Ceph
Nick Fisk - low latency CephNick Fisk - low latency Ceph
Nick Fisk - low latency Ceph
ShapeBlue
 
[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...
[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...
[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...
OpenStack Korea Community
 
Ceph as software define storage
Ceph as software define storageCeph as software define storage
Ceph as software define storage
Mahmoud Shiri Varamini
 
Ceph Introduction 2017
Ceph Introduction 2017  Ceph Introduction 2017
Ceph Introduction 2017
Karan Singh
 
2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard
Ceph Community
 
Disaggregating Ceph using NVMeoF
Disaggregating Ceph using NVMeoFDisaggregating Ceph using NVMeoF
Disaggregating Ceph using NVMeoF
ShapeBlue
 
Ceph c01
Ceph c01Ceph c01
Ceph c01
Lâm Đào
 
Multiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red Hat
Multiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red HatMultiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red Hat
Multiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red Hat
OpenStack
 
Ovs dpdk hwoffload way to full offload
Ovs dpdk hwoffload way to full offloadOvs dpdk hwoffload way to full offload
Ovs dpdk hwoffload way to full offload
Kevin Traynor
 
Ceph RBD Update - June 2021
Ceph RBD Update - June 2021Ceph RBD Update - June 2021
Ceph RBD Update - June 2021
Ceph Community
 
BlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InBlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year In
Sage Weil
 
ceph optimization on ssd ilsoo byun-short
ceph optimization on ssd ilsoo byun-shortceph optimization on ssd ilsoo byun-short
ceph optimization on ssd ilsoo byun-short
NAVER D2
 
2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph
Ceph Community
 
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake SolutionCeph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Karan Singh
 
Ceph Block Devices: A Deep Dive
Ceph Block Devices:  A Deep DiveCeph Block Devices:  A Deep Dive
Ceph Block Devices: A Deep Dive
Red_Hat_Storage
 
Ceph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing GuideCeph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing Guide
Karan Singh
 
Ceph scale testing with 10 Billion Objects
Ceph scale testing with 10 Billion ObjectsCeph scale testing with 10 Billion Objects
Ceph scale testing with 10 Billion Objects
Karan Singh
 
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA ArchitectureCeph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Danielle Womboldt
 
Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)Storage tiering and erasure coding in Ceph (SCaLE13x)
Storage tiering and erasure coding in Ceph (SCaLE13x)
Sage Weil
 
Nick Fisk - low latency Ceph
Nick Fisk - low latency CephNick Fisk - low latency Ceph
Nick Fisk - low latency Ceph
ShapeBlue
 
[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...
[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...
[OpenInfra Days Korea 2018] Day 2 - CEPH 운영자를 위한 Object Storage Performance T...
OpenStack Korea Community
 
Ceph Introduction 2017
Ceph Introduction 2017  Ceph Introduction 2017
Ceph Introduction 2017
Karan Singh
 
2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard
Ceph Community
 
Disaggregating Ceph using NVMeoF
Disaggregating Ceph using NVMeoFDisaggregating Ceph using NVMeoF
Disaggregating Ceph using NVMeoF
ShapeBlue
 
Multiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red Hat
Multiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red HatMultiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red Hat
Multiple Sites and Disaster Recovery with Ceph: Andrew Hatfield, Red Hat
OpenStack
 
Ovs dpdk hwoffload way to full offload
Ovs dpdk hwoffload way to full offloadOvs dpdk hwoffload way to full offload
Ovs dpdk hwoffload way to full offload
Kevin Traynor
 
Ceph RBD Update - June 2021
Ceph RBD Update - June 2021Ceph RBD Update - June 2021
Ceph RBD Update - June 2021
Ceph Community
 

Similar to Ceph Performance and Sizing Guide (20)

Webinar NETGEAR - ReadyNAS, le novità hardware e software
Webinar NETGEAR - ReadyNAS, le novità hardware e softwareWebinar NETGEAR - ReadyNAS, le novità hardware e software
Webinar NETGEAR - ReadyNAS, le novità hardware e software
Netgear Italia
 
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash TechnologyCeph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Ceph Community
 
JetStor NAS series 2016
JetStor NAS series 2016JetStor NAS series 2016
JetStor NAS series 2016
Gene Leyzarovich
 
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Red_Hat_Storage
 
2018 Infortrend All Flash Arrays Introduction (GS3025A)
2018 Infortrend All Flash Arrays Introduction (GS3025A)2018 Infortrend All Flash Arrays Introduction (GS3025A)
2018 Infortrend All Flash Arrays Introduction (GS3025A)
infortrendgroup
 
Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...
Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...
Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...
Principled Technologies
 
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red_Hat_Storage
 
PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...
PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...
PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...
Equnix Business Solutions
 
Accelerating hbase with nvme and bucket cache
Accelerating hbase with nvme and bucket cacheAccelerating hbase with nvme and bucket cache
Accelerating hbase with nvme and bucket cache
David Grier
 
Demystifying Storage - Building large SANs
Demystifying  Storage - Building large SANsDemystifying  Storage - Building large SANs
Demystifying Storage - Building large SANs
Directi Group
 
Open Source Data Deduplication
Open Source Data DeduplicationOpen Source Data Deduplication
Open Source Data Deduplication
RedWireServices
 
Red Hat Storage Day New York - New Reference Architectures
Red Hat Storage Day New York - New Reference ArchitecturesRed Hat Storage Day New York - New Reference Architectures
Red Hat Storage Day New York - New Reference Architectures
Red_Hat_Storage
 
Ceph Day New York 2014: Ceph, a physical perspective
Ceph Day New York 2014: Ceph, a physical perspective Ceph Day New York 2014: Ceph, a physical perspective
Ceph Day New York 2014: Ceph, a physical perspective
Ceph Community
 
Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server
Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server
Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server
Ceph Community
 
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph clusterCeph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Community
 
JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021
Gene Leyzarovich
 
Demystifying Storage
Demystifying  StorageDemystifying  Storage
Demystifying Storage
bhavintu79
 
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSAccelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Ceph Community
 
DiscoverNasbooktbs453bx01ucVlERwlR2A.pdf
DiscoverNasbooktbs453bx01ucVlERwlR2A.pdfDiscoverNasbooktbs453bx01ucVlERwlR2A.pdf
DiscoverNasbooktbs453bx01ucVlERwlR2A.pdf
nosilrub
 
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Red_Hat_Storage
 
Webinar NETGEAR - ReadyNAS, le novità hardware e software
Webinar NETGEAR - ReadyNAS, le novità hardware e softwareWebinar NETGEAR - ReadyNAS, le novità hardware e software
Webinar NETGEAR - ReadyNAS, le novità hardware e software
Netgear Italia
 
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash TechnologyCeph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Ceph Community
 
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Red_Hat_Storage
 
2018 Infortrend All Flash Arrays Introduction (GS3025A)
2018 Infortrend All Flash Arrays Introduction (GS3025A)2018 Infortrend All Flash Arrays Introduction (GS3025A)
2018 Infortrend All Flash Arrays Introduction (GS3025A)
infortrendgroup
 
Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...
Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...
Unlock more mixed storage performance on Dell PowerEdge R750 servers with Bro...
Principled Technologies
 
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red_Hat_Storage
 
PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...
PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...
PGConf.ASIA 2019 Bali - AppOS: PostgreSQL Extension for Scalable File I/O - K...
Equnix Business Solutions
 
Accelerating hbase with nvme and bucket cache
Accelerating hbase with nvme and bucket cacheAccelerating hbase with nvme and bucket cache
Accelerating hbase with nvme and bucket cache
David Grier
 
Demystifying Storage - Building large SANs
Demystifying  Storage - Building large SANsDemystifying  Storage - Building large SANs
Demystifying Storage - Building large SANs
Directi Group
 
Open Source Data Deduplication
Open Source Data DeduplicationOpen Source Data Deduplication
Open Source Data Deduplication
RedWireServices
 
Red Hat Storage Day New York - New Reference Architectures
Red Hat Storage Day New York - New Reference ArchitecturesRed Hat Storage Day New York - New Reference Architectures
Red Hat Storage Day New York - New Reference Architectures
Red_Hat_Storage
 
Ceph Day New York 2014: Ceph, a physical perspective
Ceph Day New York 2014: Ceph, a physical perspective Ceph Day New York 2014: Ceph, a physical perspective
Ceph Day New York 2014: Ceph, a physical perspective
Ceph Community
 
Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server
Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server
Ceph Day San Jose - All-Flahs Ceph on NUMA-Balanced Server
Ceph Community
 
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph clusterCeph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Community
 
JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021
Gene Leyzarovich
 
Demystifying Storage
Demystifying  StorageDemystifying  Storage
Demystifying Storage
bhavintu79
 
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSAccelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Ceph Community
 
DiscoverNasbooktbs453bx01ucVlERwlR2A.pdf
DiscoverNasbooktbs453bx01ucVlERwlR2A.pdfDiscoverNasbooktbs453bx01ucVlERwlR2A.pdf
DiscoverNasbooktbs453bx01ucVlERwlR2A.pdf
nosilrub
 
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Red_Hat_Storage
 
Ad

Recently uploaded (20)

The Significance of Hardware in Information Systems.pdf
The Significance of Hardware in Information Systems.pdfThe Significance of Hardware in Information Systems.pdf
The Significance of Hardware in Information Systems.pdf
drewplanas10
 
Adobe Master Collection CC Crack Advance Version 2025
Adobe Master Collection CC Crack Advance Version 2025Adobe Master Collection CC Crack Advance Version 2025
Adobe Master Collection CC Crack Advance Version 2025
kashifyounis067
 
Exploring Code Comprehension in Scientific Programming: Preliminary Insight...
Exploring Code Comprehension  in Scientific Programming:  Preliminary Insight...Exploring Code Comprehension  in Scientific Programming:  Preliminary Insight...
Exploring Code Comprehension in Scientific Programming: Preliminary Insight...
University of Hawai‘i at Mānoa
 
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...
Egor Kaleynik
 
How to Optimize Your AWS Environment for Improved Cloud Performance
How to Optimize Your AWS Environment for Improved Cloud PerformanceHow to Optimize Your AWS Environment for Improved Cloud Performance
How to Optimize Your AWS Environment for Improved Cloud Performance
ThousandEyes
 
Who Watches the Watchmen (SciFiDevCon 2025)
Who Watches the Watchmen (SciFiDevCon 2025)Who Watches the Watchmen (SciFiDevCon 2025)
Who Watches the Watchmen (SciFiDevCon 2025)
Allon Mureinik
 
Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)
Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)
Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)
Andre Hora
 
WinRAR Crack for Windows (100% Working 2025)
WinRAR Crack for Windows (100% Working 2025)WinRAR Crack for Windows (100% Working 2025)
WinRAR Crack for Windows (100% Working 2025)
sh607827
 
Maxon CINEMA 4D 2025 Crack FREE Download LINK
Maxon CINEMA 4D 2025 Crack FREE Download LINKMaxon CINEMA 4D 2025 Crack FREE Download LINK
Maxon CINEMA 4D 2025 Crack FREE Download LINK
younisnoman75
 
Scaling GraphRAG: Efficient Knowledge Retrieval for Enterprise AI
Scaling GraphRAG:  Efficient Knowledge Retrieval for Enterprise AIScaling GraphRAG:  Efficient Knowledge Retrieval for Enterprise AI
Scaling GraphRAG: Efficient Knowledge Retrieval for Enterprise AI
danshalev
 
Download YouTube By Click 2025 Free Full Activated
Download YouTube By Click 2025 Free Full ActivatedDownload YouTube By Click 2025 Free Full Activated
Download YouTube By Click 2025 Free Full Activated
saniamalik72555
 
Download Wondershare Filmora Crack [2025] With Latest
Download Wondershare Filmora Crack [2025] With LatestDownload Wondershare Filmora Crack [2025] With Latest
Download Wondershare Filmora Crack [2025] With Latest
tahirabibi60507
 
How can one start with crypto wallet development.pptx
How can one start with crypto wallet development.pptxHow can one start with crypto wallet development.pptx
How can one start with crypto wallet development.pptx
laravinson24
 
Automation Techniques in RPA - UiPath Certificate
Automation Techniques in RPA - UiPath CertificateAutomation Techniques in RPA - UiPath Certificate
Automation Techniques in RPA - UiPath Certificate
VICTOR MAESTRE RAMIREZ
 
Microsoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdf
Microsoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdfMicrosoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdf
Microsoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdf
TechSoup
 
Revolutionizing Residential Wi-Fi PPT.pptx
Revolutionizing Residential Wi-Fi PPT.pptxRevolutionizing Residential Wi-Fi PPT.pptx
Revolutionizing Residential Wi-Fi PPT.pptx
nidhisingh691197
 
EASEUS Partition Master Crack + License Code
EASEUS Partition Master Crack + License CodeEASEUS Partition Master Crack + License Code
EASEUS Partition Master Crack + License Code
aneelaramzan63
 
Top 10 Client Portal Software Solutions for 2025.docx
Top 10 Client Portal Software Solutions for 2025.docxTop 10 Client Portal Software Solutions for 2025.docx
Top 10 Client Portal Software Solutions for 2025.docx
Portli
 
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New Version
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New VersionPixologic ZBrush Crack Plus Activation Key [Latest 2025] New Version
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New Version
saimabibi60507
 
Get & Download Wondershare Filmora Crack Latest [2025]
Get & Download Wondershare Filmora Crack Latest [2025]Get & Download Wondershare Filmora Crack Latest [2025]
Get & Download Wondershare Filmora Crack Latest [2025]
saniaaftab72555
 
The Significance of Hardware in Information Systems.pdf
The Significance of Hardware in Information Systems.pdfThe Significance of Hardware in Information Systems.pdf
The Significance of Hardware in Information Systems.pdf
drewplanas10
 
Adobe Master Collection CC Crack Advance Version 2025
Adobe Master Collection CC Crack Advance Version 2025Adobe Master Collection CC Crack Advance Version 2025
Adobe Master Collection CC Crack Advance Version 2025
kashifyounis067
 
Exploring Code Comprehension in Scientific Programming: Preliminary Insight...
Exploring Code Comprehension  in Scientific Programming:  Preliminary Insight...Exploring Code Comprehension  in Scientific Programming:  Preliminary Insight...
Exploring Code Comprehension in Scientific Programming: Preliminary Insight...
University of Hawai‘i at Mānoa
 
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...
Egor Kaleynik
 
How to Optimize Your AWS Environment for Improved Cloud Performance
How to Optimize Your AWS Environment for Improved Cloud PerformanceHow to Optimize Your AWS Environment for Improved Cloud Performance
How to Optimize Your AWS Environment for Improved Cloud Performance
ThousandEyes
 
Who Watches the Watchmen (SciFiDevCon 2025)
Who Watches the Watchmen (SciFiDevCon 2025)Who Watches the Watchmen (SciFiDevCon 2025)
Who Watches the Watchmen (SciFiDevCon 2025)
Allon Mureinik
 
Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)
Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)
Exceptional Behaviors: How Frequently Are They Tested? (AST 2025)
Andre Hora
 
WinRAR Crack for Windows (100% Working 2025)
WinRAR Crack for Windows (100% Working 2025)WinRAR Crack for Windows (100% Working 2025)
WinRAR Crack for Windows (100% Working 2025)
sh607827
 
Maxon CINEMA 4D 2025 Crack FREE Download LINK
Maxon CINEMA 4D 2025 Crack FREE Download LINKMaxon CINEMA 4D 2025 Crack FREE Download LINK
Maxon CINEMA 4D 2025 Crack FREE Download LINK
younisnoman75
 
Scaling GraphRAG: Efficient Knowledge Retrieval for Enterprise AI
Scaling GraphRAG:  Efficient Knowledge Retrieval for Enterprise AIScaling GraphRAG:  Efficient Knowledge Retrieval for Enterprise AI
Scaling GraphRAG: Efficient Knowledge Retrieval for Enterprise AI
danshalev
 
Download YouTube By Click 2025 Free Full Activated
Download YouTube By Click 2025 Free Full ActivatedDownload YouTube By Click 2025 Free Full Activated
Download YouTube By Click 2025 Free Full Activated
saniamalik72555
 
Download Wondershare Filmora Crack [2025] With Latest
Download Wondershare Filmora Crack [2025] With LatestDownload Wondershare Filmora Crack [2025] With Latest
Download Wondershare Filmora Crack [2025] With Latest
tahirabibi60507
 
How can one start with crypto wallet development.pptx
How can one start with crypto wallet development.pptxHow can one start with crypto wallet development.pptx
How can one start with crypto wallet development.pptx
laravinson24
 
Automation Techniques in RPA - UiPath Certificate
Automation Techniques in RPA - UiPath CertificateAutomation Techniques in RPA - UiPath Certificate
Automation Techniques in RPA - UiPath Certificate
VICTOR MAESTRE RAMIREZ
 
Microsoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdf
Microsoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdfMicrosoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdf
Microsoft AI Nonprofit Use Cases and Live Demo_2025.04.30.pdf
TechSoup
 
Revolutionizing Residential Wi-Fi PPT.pptx
Revolutionizing Residential Wi-Fi PPT.pptxRevolutionizing Residential Wi-Fi PPT.pptx
Revolutionizing Residential Wi-Fi PPT.pptx
nidhisingh691197
 
EASEUS Partition Master Crack + License Code
EASEUS Partition Master Crack + License CodeEASEUS Partition Master Crack + License Code
EASEUS Partition Master Crack + License Code
aneelaramzan63
 
Top 10 Client Portal Software Solutions for 2025.docx
Top 10 Client Portal Software Solutions for 2025.docxTop 10 Client Portal Software Solutions for 2025.docx
Top 10 Client Portal Software Solutions for 2025.docx
Portli
 
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New Version
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New VersionPixologic ZBrush Crack Plus Activation Key [Latest 2025] New Version
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New Version
saimabibi60507
 
Get & Download Wondershare Filmora Crack Latest [2025]
Get & Download Wondershare Filmora Crack Latest [2025]Get & Download Wondershare Filmora Crack Latest [2025]
Get & Download Wondershare Filmora Crack Latest [2025]
saniaaftab72555
 
Ad

Ceph Performance and Sizing Guide

  • 1. Red Hat Ceph Performance & Sizing Guide Jose De la Rosa September 7, 2016
  • 2. 2 Agenda 1. Ceph Overview a. Architecture b. Approach to storing data 2. Test methodology a. What was tested b. How it was tested & measured 3. Results a. Key findings b. Recommendations
  • 4. 4 Ceph Overview 1. Open source software defined storage: logical storage services and capabilities are abstracted from the underlying physical storage systems. 2. Provides object, block and file system storage. All data is stored as objects. 3. Massively scalable to thousands of storage nodes. 4. Self-healing with no single point of failure: If a node fails, it is automatically detected and data rebalances to ensure availability.
  • 5. 5 Client access Ross Turk, Red Hat - https://raw.githubusercontent.com/ceph/ceph/master/doc/images/stack.png
  • 7. 7 Storage Protection Method Disk Disk Disk OSD OSD OSD 3x Replication Data copied 3 times Actual disk usage = 33% Disk Disk Disk Disk OSD OSD OSD OSD Erasure Coded (3+2) Data split into 3 + 2 disks used for parity Actual disk usage = 60% Disk OSD
  • 8. 8 Reading and Writing Data http://www.slideshare.net/LarryCover/ceph-open-source-storage-software-optimizations-on-intel-architecture-for-cloud-workloads
  • 10. 10 Server configuration Testbed Details Ceph tier Storage Nodes (5) Monitors (3) Clients (10) Platform Dell PowerEdge R730xd Dell PowerEdge R630 Dell PowerEdge R220 CPU 2x Intel Xeon E5-2630 v3 2.4GHz 2x Intel Xeon E5-2650 v3 2.3 GHz 1x Intel Celeron G1820 2.7 GHz Memory 4x 16 GB 1866 MHz DDR4 8x 16 GB 2133MHz DDR4 4x 4 GB 1600 MHz DDR3 Network 1x Intel X520/2P I350 LOM 1x Intel X520/2P I350 LOM 1x Intel X520/2P I350 Storage PERC H730 Mini / 1 GB Cache Up to 16x: SEAGATE 4 TB SAS (ST4000NM0005) Up to 3x: Intel DC S3700 SSD 200 GB SATA (SSDSC2BA20) 1x Intel DC P3700 SSD 800 GB NVMe PERC H730 Mini / 1 GB Cache 6x SEAGATE 500 GB SAS (ST9500620S) 1x Toshiba 50 GB SATA (DT01ACA0)
  • 12. 12 Configuration Guidance 1. General rules of thumb • 1 Core-GHz per OSD • SATA/SAS SSD to HDD Ratio: 1:4 - 1:5 • NVME SSD to HDD Ratio: 1:17-1:18 • 16GB RAM Baseline + 2-3GB per OSD 2. More details at https://www.redhat.com/en/resources/red-hat- ceph-storage-hardware-configuration-guide
  • 13. 13 Storage node configurations tested OSD to Journal Ratio [drives] 12+3 16+0 16+1 OSD node configuration 12+3 16+0 16+1 HDDs 12 16 16 HDD RAID mode Single-disk RAID0 Single-disk RAID0 Single-disk RAID0 / HBA mode SATA SSDs 3 0 0 SSD RAID mode JBOD JBOD JBOD NVMe SSDs 0 0 1 Network 1x 10 GbE Front-End 1x 10 GbE Back-End 1x 10 GbE Front-End 1x 10 GbE Back-End 1x 10 GbE Front-End 1x 10 GbE Back-End
  • 14. 14 Benchmarking with CBT 1. For benchmark automation, the open source utility Ceph Benchmarking Tool (CBT) was used. 2. It supports different drivers for examining different layers of the storage stack: • radosbench - uses librados API (used in this study) • librbdfio – test block storage without KVM/QEMU instances • kvmrbdfio – test block volumes attached to KVM/QEMU instances 3. Available at https://github.com/ceph/cbt
  • 15. 15 Factors that influence performance 1. Device used for journaling (SSD vs. HDD) 2. RAID0 vs. pass-through (HBA) mode 3. Number of clients (single stream vs. parallel access) 4. Data-protection mode (3-way replication vs. erasure coding)
  • 16. 16 Test Matrix Server configuration PowerEdge R730xd 12+3, 3xRep PowerEdge R730xd 16+0, EC3+2 PowerEdge R730xd 16r+1, 3xRep PowerEdge R730xd 16+1, EC 3+2 PowerEdge R730xd 16j+1, 3xRep OS disk 2x 500 GB 2.5" 2x 500 GB 2.5" 2x 500 GB 2.5" 2x 500 GB 2.5" 2x 500 GB 2.5" Data disk type HDD 7.2K SAS 12Gbps, 4TB HDD 7.2K SAS 12Gbps, 4TB HDD 7.2K SAS 12Gbps, 4TB HDD 7.2K SAS 12Gbps, 4TB HDD 7.2K SAS 12Gbps, 4TB HDD quantity 12 16 16 16 16 Number of Ceph write journal devices 3 0 1 1 1 Ceph write journal device type Intel SATA SSD S3710 (6Gb/s) n/a Intel P3700 PCIe NVMe HHHL AIC Intel P3700 PCIe NVMe HHHL AIC Intel P3700 PCIe NVMe HHHL AIC Ceph write journal device size (GB) 200 0 800 800 800 Controller model PERC H730, 1 GB Cache PERC H730, 1 GB Cache PERC H730, 1 GB Cache PERC H730, 1 GB Cache PERC H730, 1 GB Cache PERC Controller configuration for HDDs RAID RAID RAID RAID JBOD (PERC pass-through mode) Raw capacity for Ceph OSDs (TB) 48 64 64 64 64
  • 18. 18 Throughput / server 0 200 400 600 800 1000 1200 1400 R730xd 16r+1, 3xRep R730xd 16j+1, 3xRep R730XD 16+1, EC3+2 R730xd 16+0, EC3+2 R730xd 12+3, 3xRep MBps per Server (4MB seq IO) Performance Comparison Throughput per Server Writes Reads
  • 19. 19 Overall Solution Price/Performance R730xd 16+0, EC3+2 R730xd 16+0, EC8+3 R730xd 16r+1, 3xRep R730xd 16j+1, 3xRep R730XD 16+1, EC3+2 R730xd 16+1, EC8+3 R730xd 12+3, 3xRep Total Cluster Server+SW Cost / Cluster MBps Solution Price/Performance Comparison 500TB Usable Cluster (less $ per MBps is better) Write Read $0
  • 20. 20 Overall Solution Price Capacity R730xd 16+0, EC3+2 R730xd 16+0, EC8+3 R730xd 16r+1, 3xRep R730xd 16j+1, 3xRep R730XD 16+1, EC3+2 R730xd 16+1, EC8+3 R730xd 12+3, 3xRep Total Cluster Server+SW Cost / Cluster GB Solution Price/Capacity Comparison (less $ per GB is better) $0
  • 21. 21 Replication vs. Erasure Coding 0 200 400 600 800 1000 1200 1400 R730xd 16r+1, 3xRep R730xd 16j+1, 3xRep R730XD 16+1, EC3+2 R730xd 16+1, EC8+3 MBps per Server (4MB seq IO) Performance Comparison Replication vs. Erasure-coding Writes Reads
  • 22. 22 JBOD vs. RAID0 1153 1147 393 342 0 200 400 600 800 1000 1200 1400 R730XD 16 RAID+1, 3xRep R730XD 16 JBOD+1, 3xRep MBps per Server (4MB seq IO) Performance Comparison JBOD vs. RAID0 Config Writes Reads
  • 23. 23 Performance conclusions 1. Replication mode yielded better performance for read operations and the erasure-coded mode proved better for write operations. 2. The PowerEdge R730xd 16+1 3x replication configuration yielded optimal price for read-write throughput-oriented workloads. 3. The PowerEdge R730xd 12+3 3x replication configuration yielded optimal price for read-only throughput-oriented workloads. 4. The PowerEdge R730xd 16+1 erasure-coded configuration proved to be the choice for write-heavy operations. 5. When used with Ceph Storage, Dell & Red Hat recommend the usage of single-drive RAID0 mode on PowerEdge R730xd with PERC H730.
  • 24. 24 Sizing Recommendations Storage Capacity Extra Small Small Medium Cluster Capacity 100 TB+ 500 TB+ 1 PB+ Throughput-Optimized >4x R730xd (8U) >8x R730xd (16U) NA 1x server/2U chassis 1x server/2U chassis 16x 6 TB HDD 16x 6 TB HDD 1x 800 GB NVMe SSD 1x 800 GB NVMe SSD 2x 10 GbE 2x 10 GbE 3x Replication 3x Replication Cost/Capacity-Optimized NA NA >15x R730xd (30U) 1x server/2U chassis 16x 8 TB HDD 1x HHHL AIC SSD 2x 10 GbE 8:3 Erasure-coding
  • 25. 25 Observations 1. Obey SSD to HDD ratio 2. Hardware matters, look at RAID controllers if you use HDDs 3. Don’t use RAID controllers on SSDs 4. SSD sequential write bandwidth becomes a bottleneck 5. Random workloads should go on Flash-only 6. 10GbE Bonding not necessary with <=16 drives
  • 26. 26 Recommended Reading Dell PowerEdge R730xd Red Hat Ceph Storage Performance and Sizing Guide http://en.community.dell.com/techcenter/cloud/m/dell_cloud_resources/20442913 Enterprise Ceph: Every Way, Your Way https://www.redhat.com/files/summit/session-assets/2016/SS88828-enterprise- ceph_every-way-your-way.pdf

Editor's Notes

  • #6: http://docs.ceph.com/docs/hammer/architecture/ RADOS - preserves consistent data access and strong safety semantics while allowing nodes to act semi-autonomously to self manage replication, failure detection, and failure recovery through the use of a small cluster map.
  • #7: Monitor nodes – which monitor the state of the cluster and decide what storage nodes are available to store data Metadata servers store metadata for CephFS file systems Data is always evenly distributed to all OSDs CRUSH is configurable and scales indefinitely
  • #8: OSDs perform optimally when all storage drives are the same size and speed. Solid state storage media (SSDs and NVMe) can be used for journaling and caching to improve performance + consistency.
  • #14: Throughput is a measurement of the average number of megabytes transferred within a period of time for a specific file size.
  • #19: The R730xd 16+1, 3xRep configuration provided the best performance for read/write workloads. However, for write-oriented workloads, the R730xd 16+1, EC3+2 configuration provides superior write performance at significantly lesser costs.
  • #20: The R730xd 16+1, 3xRep configuration yielded optimal price performance for mixed read/write, and throughput-oriented workloads. However, for read-mostly workloads, the R730xd 12+3, 3xRep configuration is an attractive alternative based on its superior read price/performance.
  • #21: For capacity-archive workloads, erasure-coded configurations are significantly less expensive per GB data archived. Write-heavy capacity-archives should use the R730xd 16+1, EC configuration, because adding an SSD write-journal increases total $/GB by only a small value and increases write performance.
  • #22: Keeping everything else constant, replicated reads perform much better than erasure-coded reads. However, erasure-coded writes perform better than replicated writes.