SlideShare a Scribd company logo
1 of 24
Download to read offline
HA NAS with CephFSWyllys Ingersoll – Keeper Technology, LLC
Data and Storage Management Experts
Focus on IC & Commercial Customers for 12 Years
•  Mul?-PB Enterprise Systems
•  Imagery, Computer Forensics, Big Data
•  High Volume/High Velocity Data Analysis
•  Full Solu?on Provider
•  Keeper Products + Partner Products

Introductions
Keeper Technology
•  How we implemented HA NAS gateways using
cephfs
•  Cluster configura?on
•  SoPware used
•  Issues encountered
•  Performance Sta?s?cs
Overview
Overview
3/17/17 - 3
•  Ceph Jewel 10.2.5 on Ubuntu-based Linux
•  6 Storage servers, ~ 80 TB Usable (3-copy)
•  ~15 OSD per server
•  2, 3, and 4TB 7200RPM spinning drives (no SSD)
•  2 Gateways
•  HP DL380 G9 Servers w/48GB RAM
•  3 Monitors + 3 MDS Servers
•  HP DL360 G6 w/48GB RAM
Configuration
Cluster Configuration
3/17/17 - 4
•  Provide NFS and/or SMB filesystem shares with
redundancy
•  Failover if a gateway goes down
•  Clients should not lose data
•  Minimal interrup?on of client workflow
•  “Seamless” for NFSv3 – others are WiP.
•  Minimum 2 gateways required
Goals
HA NAS Goals
3/17/17 - 5
•  SAMBA 4.5.5 w/CTDB support
•  Built with “—with-cluster-support” flag
•  CTDB is key to HA func?onality
•  CTDB = Clustered Trivial Database
•  Node monitor, failover, IP takeover
•  Define mul?ple floa?ng IP Addresses in DNS
•  CTDB configured with virtual shared IPs and real IP of
each GW.
•  CTDB nodes communicate on private network
•  Insecure protocol
Software
SMB
3/17/17 - 6
•  Ganesha NFS 2.4.3
•  User space NFS service, replaces kernel NFS
•  Building code from github repo code
•  Store ganesha config on shared FS
•  Ex: /cephfs/nfs/ganesha.conf
•  HA gateways must have common NFS Export IDs
•  Use “VFS” FSAL (not “CEPH”) for Ganesha exports
Software
NFS
3/17/17 - 7
•  Single FS per cluster (for now) - /cephfs
•  Disable snapshots or restrict to top level
•  Hard-linking bug prevents reliable snapshots on subdirs
•  Prefer kernel mounted over fuse for performance
•  Kernel 4.8.10
•  Each export is a subdirectory
•  /cephfs/exports/foobar
Software
NAS with Cephfs
3/17/17 - 8
Data Flow
3/17/17 - 9
10GB
Ceph
GW-1
GW-2
GW-3
Virtual IP
Client
1
Client
2
Client
3
SMB
CTDB
CTDB
CTDB
10GB Network
•  SAMBA locks stored on shared FS
•  Ex: /cephfs/ctdb
•  CTDB monitors SAMBA and Ganesha services
•  Starts and stops as necessary via “callout” scripts
•  CTDB assigns virtual IP addrs as needed
Software
NAS With Cephfs
3/17/17 - 10
•  Kernel support for cephfs varies
•  Using “bleeding edge” kernels for best results
•  Cannot set quotas on subdirectories
•  kernel cephfs limita?on
•  Cannot limit size available for a single export
•  Each share has max size = en?re cephfs data pool
•  Snapshots only at top level
•  Cannot snapshot each exported subdirectory
Issues
Issues and Problems
3/17/17 - 11
•  mds_cache_size = 8,000,000 (default was 100k)
•  Uses more RAM, but we have 48GB
•  Avoid “failing to respond to cache pressure” errors
•  Use “default” crush tunables (not “jewel”).
•  Works beqer with older kernels

Issues
Adjustments
3/17/17 - 12
•  FIO parameters
•  Vary block sizes (4K, 64K, 1m, 4m)
•  Vary # of jobs (1, 16, 32, 64)
•  Iodepth = 1
•  Read/write + randread/randwrite
•  Ioengine = sync
•  direct=1
•  Simultaneous 4 dis?nct clients on 10GB link
Performance
Test Methodology
3/17/17 - 13
Performance
Performance Configuration
3/17/17 - 14
NAS Gateway
Client
1
Client
2
Client
3
Client
4
/cephfs
Shared FS (/cephfs)
mounted with NFS or SMB
/cephfs
 /cephfs
 /cephfs
 /cephfs
FIO directIO used on each client
to read & write data to the share.
Ceph Cluster
Performance
NFS Read (kernel mount)
3/17/17 - 15
Performance
NFS Writes (kernel mount)
3/17/17 - 16
Performance
NFS Read (Fuse mount)
3/17/17 - 17
Performance
NFS Write (Fuse mount)
3/17/17 - 18
Performance
SMB Read (kernel mount)
3/17/17 - 19
Performance
SMB Write (kernel mount)
3/17/17 - 20
Performance
SMB Read (fuse mount)
3/17/17 - 21
Performance
SMB Write (fuse mount)
3/17/17 - 22
•  High Availability NAS is possible with cephfs
•  Some issues remain
•  Deep snapshots
•  Quotas/limits on subdirs
•  Performance is “OK”
•  Standard NAS protocol limita?ons (NFS, SMB)
Summary
Summary
3/17/17 - 23
Thank You 
| 21740 Beaumeade Circle | Suite 150 | Ashburn, VA 20147 | P [571] 333 2725 | F [703] 738 7231 | solu?ons@keepertech.com | www.keepertech.com
Thank You

More Related Content

What's hot

Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
Ceph Community
 
Ceph Day Taipei - How ARM Microserver Cluster Performs in Ceph
Ceph Day Taipei - How ARM Microserver Cluster Performs in CephCeph Day Taipei - How ARM Microserver Cluster Performs in Ceph
Ceph Day Taipei - How ARM Microserver Cluster Performs in Ceph
Ceph Community
 
Ceph Day Shanghai - Recovery Erasure Coding and Cache Tiering
Ceph Day Shanghai - Recovery Erasure Coding and Cache TieringCeph Day Shanghai - Recovery Erasure Coding and Cache Tiering
Ceph Day Shanghai - Recovery Erasure Coding and Cache Tiering
Ceph Community
 

What's hot (20)

Ceph Day KL - Ceph on All-Flash Storage
Ceph Day KL - Ceph on All-Flash Storage Ceph Day KL - Ceph on All-Flash Storage
Ceph Day KL - Ceph on All-Flash Storage
 
2016-JAN-28 -- High Performance Production Databases on Ceph
2016-JAN-28 -- High Performance Production Databases on Ceph2016-JAN-28 -- High Performance Production Databases on Ceph
2016-JAN-28 -- High Performance Production Databases on Ceph
 
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
 
Ceph Day KL - Delivering cost-effective, high performance Ceph cluster
Ceph Day KL - Delivering cost-effective, high performance Ceph clusterCeph Day KL - Delivering cost-effective, high performance Ceph cluster
Ceph Day KL - Delivering cost-effective, high performance Ceph cluster
 
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA ArchitectureCeph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
 
Ceph Tech Talk -- Ceph Benchmarking Tool
Ceph Tech Talk -- Ceph Benchmarking ToolCeph Tech Talk -- Ceph Benchmarking Tool
Ceph Tech Talk -- Ceph Benchmarking Tool
 
Ceph Performance Profiling and Reporting
Ceph Performance Profiling and ReportingCeph Performance Profiling and Reporting
Ceph Performance Profiling and Reporting
 
Ceph Day Taipei - How ARM Microserver Cluster Performs in Ceph
Ceph Day Taipei - How ARM Microserver Cluster Performs in CephCeph Day Taipei - How ARM Microserver Cluster Performs in Ceph
Ceph Day Taipei - How ARM Microserver Cluster Performs in Ceph
 
Ceph Day Shanghai - Recovery Erasure Coding and Cache Tiering
Ceph Day Shanghai - Recovery Erasure Coding and Cache TieringCeph Day Shanghai - Recovery Erasure Coding and Cache Tiering
Ceph Day Shanghai - Recovery Erasure Coding and Cache Tiering
 
Ceph on All Flash Storage -- Breaking Performance Barriers
Ceph on All Flash Storage -- Breaking Performance BarriersCeph on All Flash Storage -- Breaking Performance Barriers
Ceph on All Flash Storage -- Breaking Performance Barriers
 
Ceph Day Tokyo - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise Ceph Day Tokyo - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise
 
MySQL Head-to-Head
MySQL Head-to-HeadMySQL Head-to-Head
MySQL Head-to-Head
 
Ceph Day Taipei - Accelerate Ceph via SPDK
Ceph Day Taipei - Accelerate Ceph via SPDK Ceph Day Taipei - Accelerate Ceph via SPDK
Ceph Day Taipei - Accelerate Ceph via SPDK
 
Ceph Day Taipei - Delivering cost-effective, high performance, Ceph cluster
Ceph Day Taipei - Delivering cost-effective, high performance, Ceph cluster Ceph Day Taipei - Delivering cost-effective, high performance, Ceph cluster
Ceph Day Taipei - Delivering cost-effective, high performance, Ceph cluster
 
Ceph on 64-bit ARM with X-Gene
Ceph on 64-bit ARM with X-GeneCeph on 64-bit ARM with X-Gene
Ceph on 64-bit ARM with X-Gene
 
QCT Ceph Solution - Design Consideration and Reference Architecture
QCT Ceph Solution - Design Consideration and Reference ArchitectureQCT Ceph Solution - Design Consideration and Reference Architecture
QCT Ceph Solution - Design Consideration and Reference Architecture
 
Ceph Day Beijing - Ceph RDMA Update
Ceph Day Beijing - Ceph RDMA UpdateCeph Day Beijing - Ceph RDMA Update
Ceph Day Beijing - Ceph RDMA Update
 
Developing a Ceph Appliance for Secure Environments
Developing a Ceph Appliance for Secure EnvironmentsDeveloping a Ceph Appliance for Secure Environments
Developing a Ceph Appliance for Secure Environments
 
Ceph: Low Fail Go Scale
Ceph: Low Fail Go Scale Ceph: Low Fail Go Scale
Ceph: Low Fail Go Scale
 
Ceph Day Melbourne - Ceph on All-Flash Storage - Breaking Performance Barriers
Ceph Day Melbourne - Ceph on All-Flash Storage - Breaking Performance BarriersCeph Day Melbourne - Ceph on All-Flash Storage - Breaking Performance Barriers
Ceph Day Melbourne - Ceph on All-Flash Storage - Breaking Performance Barriers
 

Viewers also liked

Web security-–-everything-we-know-is-wrong-eoin-keary
Web security-–-everything-we-know-is-wrong-eoin-kearyWeb security-–-everything-we-know-is-wrong-eoin-keary
Web security-–-everything-we-know-is-wrong-eoin-keary
drewz lin
 

Viewers also liked (20)

Ceph Day San Jose - Ceph in a Post-Cloud World
Ceph Day San Jose - Ceph in a Post-Cloud World Ceph Day San Jose - Ceph in a Post-Cloud World
Ceph Day San Jose - Ceph in a Post-Cloud World
 
Ceph Day Tokyo - Ceph Community Update
Ceph Day Tokyo - Ceph Community Update Ceph Day Tokyo - Ceph Community Update
Ceph Day Tokyo - Ceph Community Update
 
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
Ceph Day Tokyo - Ceph on ARM: Scaleable and Efficient
 
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
Ceph Day Tokyo - Bit-Isle's 3 years footprint with Ceph
 
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph clusterCeph Day Tokyo - Delivering cost effective, high performance Ceph cluster
Ceph Day Tokyo - Delivering cost effective, high performance Ceph cluster
 
Ceph Day Seoul - AFCeph: SKT Scale Out Storage Ceph
Ceph Day Seoul - AFCeph: SKT Scale Out Storage Ceph Ceph Day Seoul - AFCeph: SKT Scale Out Storage Ceph
Ceph Day Seoul - AFCeph: SKT Scale Out Storage Ceph
 
Ceph Day Seoul - Delivering Cost Effective, High Performance Ceph cluster
Ceph Day Seoul - Delivering Cost Effective, High Performance Ceph cluster Ceph Day Seoul - Delivering Cost Effective, High Performance Ceph cluster
Ceph Day Seoul - Delivering Cost Effective, High Performance Ceph cluster
 
Ceph Day Seoul - Ceph on All-Flash Storage
Ceph Day Seoul - Ceph on All-Flash Storage Ceph Day Seoul - Ceph on All-Flash Storage
Ceph Day Seoul - Ceph on All-Flash Storage
 
Ceph Day Tokyo - High Performance Layered Architecture
Ceph Day Tokyo - High Performance Layered Architecture  Ceph Day Tokyo - High Performance Layered Architecture
Ceph Day Tokyo - High Performance Layered Architecture
 
Ceph Day Seoul - Community Update
Ceph Day Seoul - Community UpdateCeph Day Seoul - Community Update
Ceph Day Seoul - Community Update
 
Red Hat Storage Day Dallas - Storage for OpenShift Containers
Red Hat Storage Day Dallas - Storage for OpenShift Containers Red Hat Storage Day Dallas - Storage for OpenShift Containers
Red Hat Storage Day Dallas - Storage for OpenShift Containers
 
London Ceph Day: Unified Cloud Storage with Synnefo + Ceph + Ganeti
London Ceph Day: Unified Cloud Storage with Synnefo + Ceph + GanetiLondon Ceph Day: Unified Cloud Storage with Synnefo + Ceph + Ganeti
London Ceph Day: Unified Cloud Storage with Synnefo + Ceph + Ganeti
 
DataEngConf: Apache Kafka at Rocana: a scalable, distributed log for machine ...
DataEngConf: Apache Kafka at Rocana: a scalable, distributed log for machine ...DataEngConf: Apache Kafka at Rocana: a scalable, distributed log for machine ...
DataEngConf: Apache Kafka at Rocana: a scalable, distributed log for machine ...
 
Performance Metrics and Ontology for Describing Performance Data of Grid Work...
Performance Metrics and Ontology for Describing Performance Data of Grid Work...Performance Metrics and Ontology for Describing Performance Data of Grid Work...
Performance Metrics and Ontology for Describing Performance Data of Grid Work...
 
Web security-–-everything-we-know-is-wrong-eoin-keary
Web security-–-everything-we-know-is-wrong-eoin-kearyWeb security-–-everything-we-know-is-wrong-eoin-keary
Web security-–-everything-we-know-is-wrong-eoin-keary
 
Ford
FordFord
Ford
 
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
Ceph, Now and Later: Our Plan for Open Unified Cloud StorageCeph, Now and Later: Our Plan for Open Unified Cloud Storage
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
 
Connected Vehicle Data Platform
Connected Vehicle Data PlatformConnected Vehicle Data Platform
Connected Vehicle Data Platform
 
Double Your Hadoop Hardware Performance with SmartSense
Double Your Hadoop Hardware Performance with SmartSenseDouble Your Hadoop Hardware Performance with SmartSense
Double Your Hadoop Hardware Performance with SmartSense
 
HPC Top 5 Stories: March 22, 2017
HPC Top 5 Stories: March 22, 2017HPC Top 5 Stories: March 22, 2017
HPC Top 5 Stories: March 22, 2017
 

Similar to Ceph Day San Jose - HA NAS with CephFS

Deploying flash storage for Ceph without compromising performance
Deploying flash storage for Ceph without compromising performance Deploying flash storage for Ceph without compromising performance
Deploying flash storage for Ceph without compromising performance
Ceph Community
 
Introduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3RIntroduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3R
Simon Huang
 
SOUG_GV_Flashgrid_V4
SOUG_GV_Flashgrid_V4SOUG_GV_Flashgrid_V4
SOUG_GV_Flashgrid_V4
UniFabric
 
[B4]deview 2012-hdfs
[B4]deview 2012-hdfs[B4]deview 2012-hdfs
[B4]deview 2012-hdfs
NAVER D2
 

Similar to Ceph Day San Jose - HA NAS with CephFS (20)

Red Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed Hat Gluster Storage Performance
Red Hat Gluster Storage Performance
 
DPDK Summit 2015 - Aspera - Charles Shiflett
DPDK Summit 2015 - Aspera - Charles ShiflettDPDK Summit 2015 - Aspera - Charles Shiflett
DPDK Summit 2015 - Aspera - Charles Shiflett
 
Hive spark-s3acommitter-hbase-nfs
Hive spark-s3acommitter-hbase-nfsHive spark-s3acommitter-hbase-nfs
Hive spark-s3acommitter-hbase-nfs
 
Deploying flash storage for Ceph without compromising performance
Deploying flash storage for Ceph without compromising performance Deploying flash storage for Ceph without compromising performance
Deploying flash storage for Ceph without compromising performance
 
Spectrum Scale Unified File and Object with WAN Caching
Spectrum Scale Unified File and Object with WAN CachingSpectrum Scale Unified File and Object with WAN Caching
Spectrum Scale Unified File and Object with WAN Caching
 
Software Defined Analytics with File and Object Access Plus Geographically Di...
Software Defined Analytics with File and Object Access Plus Geographically Di...Software Defined Analytics with File and Object Access Plus Geographically Di...
Software Defined Analytics with File and Object Access Plus Geographically Di...
 
Performance characterization in large distributed file system with gluster fs
Performance characterization in large distributed file system with gluster fsPerformance characterization in large distributed file system with gluster fs
Performance characterization in large distributed file system with gluster fs
 
Introduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3RIntroduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3R
 
QNAP TS-832PX-4G.pdf
QNAP TS-832PX-4G.pdfQNAP TS-832PX-4G.pdf
QNAP TS-832PX-4G.pdf
 
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...Disaggregated Container Attached Storage - Yet Another Topology with What Pur...
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...
 
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...Disaggregated Container Attached Storage - Yet Another Topology with What Pur...
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...
 
What's new in Hadoop Common and HDFS
What's new in Hadoop Common and HDFS What's new in Hadoop Common and HDFS
What's new in Hadoop Common and HDFS
 
Red Hat Ceph Storage Acceleration Utilizing Flash Technology
Red Hat Ceph Storage Acceleration Utilizing Flash Technology Red Hat Ceph Storage Acceleration Utilizing Flash Technology
Red Hat Ceph Storage Acceleration Utilizing Flash Technology
 
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
 
SOUG_GV_Flashgrid_V4
SOUG_GV_Flashgrid_V4SOUG_GV_Flashgrid_V4
SOUG_GV_Flashgrid_V4
 
Deploying pNFS over Distributed File Storage w/ Jiffin Tony Thottan and Niels...
Deploying pNFS over Distributed File Storage w/ Jiffin Tony Thottan and Niels...Deploying pNFS over Distributed File Storage w/ Jiffin Tony Thottan and Niels...
Deploying pNFS over Distributed File Storage w/ Jiffin Tony Thottan and Niels...
 
Application Caching: The Hidden Microservice
Application Caching: The Hidden MicroserviceApplication Caching: The Hidden Microservice
Application Caching: The Hidden Microservice
 
Born to be fast! - Aviram Bar Haim - OpenStack Israel 2017
Born to be fast! - Aviram Bar Haim - OpenStack Israel 2017Born to be fast! - Aviram Bar Haim - OpenStack Israel 2017
Born to be fast! - Aviram Bar Haim - OpenStack Israel 2017
 
[B4]deview 2012-hdfs
[B4]deview 2012-hdfs[B4]deview 2012-hdfs
[B4]deview 2012-hdfs
 
Storage spaces direct webinar
Storage spaces direct webinarStorage spaces direct webinar
Storage spaces direct webinar
 

Recently uploaded

Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
vu2urc
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
Enterprise Knowledge
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
Earley Information Science
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
giselly40
 

Recently uploaded (20)

🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 

Ceph Day San Jose - HA NAS with CephFS

  • 1. HA NAS with CephFSWyllys Ingersoll – Keeper Technology, LLC
  • 2. Data and Storage Management Experts Focus on IC & Commercial Customers for 12 Years •  Mul?-PB Enterprise Systems •  Imagery, Computer Forensics, Big Data •  High Volume/High Velocity Data Analysis •  Full Solu?on Provider •  Keeper Products + Partner Products Introductions Keeper Technology
  • 3. •  How we implemented HA NAS gateways using cephfs •  Cluster configura?on •  SoPware used •  Issues encountered •  Performance Sta?s?cs Overview Overview 3/17/17 - 3
  • 4. •  Ceph Jewel 10.2.5 on Ubuntu-based Linux •  6 Storage servers, ~ 80 TB Usable (3-copy) •  ~15 OSD per server •  2, 3, and 4TB 7200RPM spinning drives (no SSD) •  2 Gateways •  HP DL380 G9 Servers w/48GB RAM •  3 Monitors + 3 MDS Servers •  HP DL360 G6 w/48GB RAM Configuration Cluster Configuration 3/17/17 - 4
  • 5. •  Provide NFS and/or SMB filesystem shares with redundancy •  Failover if a gateway goes down •  Clients should not lose data •  Minimal interrup?on of client workflow •  “Seamless” for NFSv3 – others are WiP. •  Minimum 2 gateways required Goals HA NAS Goals 3/17/17 - 5
  • 6. •  SAMBA 4.5.5 w/CTDB support •  Built with “—with-cluster-support” flag •  CTDB is key to HA func?onality •  CTDB = Clustered Trivial Database •  Node monitor, failover, IP takeover •  Define mul?ple floa?ng IP Addresses in DNS •  CTDB configured with virtual shared IPs and real IP of each GW. •  CTDB nodes communicate on private network •  Insecure protocol Software SMB 3/17/17 - 6
  • 7. •  Ganesha NFS 2.4.3 •  User space NFS service, replaces kernel NFS •  Building code from github repo code •  Store ganesha config on shared FS •  Ex: /cephfs/nfs/ganesha.conf •  HA gateways must have common NFS Export IDs •  Use “VFS” FSAL (not “CEPH”) for Ganesha exports Software NFS 3/17/17 - 7
  • 8. •  Single FS per cluster (for now) - /cephfs •  Disable snapshots or restrict to top level •  Hard-linking bug prevents reliable snapshots on subdirs •  Prefer kernel mounted over fuse for performance •  Kernel 4.8.10 •  Each export is a subdirectory •  /cephfs/exports/foobar Software NAS with Cephfs 3/17/17 - 8
  • 9. Data Flow 3/17/17 - 9 10GB Ceph GW-1 GW-2 GW-3 Virtual IP Client 1 Client 2 Client 3 SMB CTDB CTDB CTDB 10GB Network
  • 10. •  SAMBA locks stored on shared FS •  Ex: /cephfs/ctdb •  CTDB monitors SAMBA and Ganesha services •  Starts and stops as necessary via “callout” scripts •  CTDB assigns virtual IP addrs as needed Software NAS With Cephfs 3/17/17 - 10
  • 11. •  Kernel support for cephfs varies •  Using “bleeding edge” kernels for best results •  Cannot set quotas on subdirectories •  kernel cephfs limita?on •  Cannot limit size available for a single export •  Each share has max size = en?re cephfs data pool •  Snapshots only at top level •  Cannot snapshot each exported subdirectory Issues Issues and Problems 3/17/17 - 11
  • 12. •  mds_cache_size = 8,000,000 (default was 100k) •  Uses more RAM, but we have 48GB •  Avoid “failing to respond to cache pressure” errors •  Use “default” crush tunables (not “jewel”). •  Works beqer with older kernels Issues Adjustments 3/17/17 - 12
  • 13. •  FIO parameters •  Vary block sizes (4K, 64K, 1m, 4m) •  Vary # of jobs (1, 16, 32, 64) •  Iodepth = 1 •  Read/write + randread/randwrite •  Ioengine = sync •  direct=1 •  Simultaneous 4 dis?nct clients on 10GB link Performance Test Methodology 3/17/17 - 13
  • 14. Performance Performance Configuration 3/17/17 - 14 NAS Gateway Client 1 Client 2 Client 3 Client 4 /cephfs Shared FS (/cephfs) mounted with NFS or SMB /cephfs /cephfs /cephfs /cephfs FIO directIO used on each client to read & write data to the share. Ceph Cluster
  • 15. Performance NFS Read (kernel mount) 3/17/17 - 15
  • 16. Performance NFS Writes (kernel mount) 3/17/17 - 16
  • 17. Performance NFS Read (Fuse mount) 3/17/17 - 17
  • 18. Performance NFS Write (Fuse mount) 3/17/17 - 18
  • 19. Performance SMB Read (kernel mount) 3/17/17 - 19
  • 20. Performance SMB Write (kernel mount) 3/17/17 - 20
  • 21. Performance SMB Read (fuse mount) 3/17/17 - 21
  • 22. Performance SMB Write (fuse mount) 3/17/17 - 22
  • 23. •  High Availability NAS is possible with cephfs •  Some issues remain •  Deep snapshots •  Quotas/limits on subdirs •  Performance is “OK” •  Standard NAS protocol limita?ons (NFS, SMB) Summary Summary 3/17/17 - 23
  • 24. Thank You | 21740 Beaumeade Circle | Suite 150 | Ashburn, VA 20147 | P [571] 333 2725 | F [703] 738 7231 | solu?ons@keepertech.com | www.keepertech.com Thank You