All the distributed file systems and NoSQL databases work very well during the normal operation. We can find the big differences if we investigate the behaviour in case of emergency. Data replication strategies and recovery algorithms are the key ingredients of a distributed data storage to save data.
Recent studies proved that the random replication is not the safest choice for storing data as it almost guarantees to lose data in the common scenario of simultaneous node failures. Copyset Replication method significantly reduces the frequency of data loss events with selecting the replica groups in a smart way.
In this talk we will introduce the key elements of a successful data replication and show how advanced data replication strategies could help to survive outages. We will show how Apache Hadoop Ozone solves the problem with advanced techniques and present the challenges of using Copyset algorithm with advanced cluster topology support.
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
Losing Data in a Safe Way – Advanced Replication Strategies in Apache Hadoop Ozone
1. Losing Data in a Safe WayLosing Data in a Safe Way
Advanced Replication Strategies in ApacheAdvanced Replication Strategies in Apache
Hadoop OzoneHadoop Ozone
7. Storage in the worldStorage in the world
hadoop-hdfs hadoop-aws/aliyun/azure
8. Storage in the worldStorage in the world
hadoop-hdfs hadoop-aws/aliyun/azure
(Eventual) Consistency
Cost (PB scale)
Slower (no local data)
Easy to use / no management
Can't handle
small files
Tool support
Only HadoopFS
protocol is supported
11. Replication: split the filesReplication: split the files
Namenode Datanode
Report the state of all
the managed blocks
In-memory
File -> []block mapping
20. What is Ozone?What is Ozone?
Ozone provides Object Store
semantic (S3) for Hadoop storage
HDDS: On top of a lower level
replication layer (Hadoop Distributed
Data Store)
Consistent
Fast
Cloud native
Easy to use
21. Bonus: reporting superblocksBonus: reporting superblocks
Replicate blocks in bigger groups:
Can handle 10 billions of blocks
Namenode Datanode
22. Bonus: reporting superblocksBonus: reporting superblocks
Replicate blocks in bigger groups:
Can handle 10 billions of blocks
Namenode Datanode
23. Open containersOpen containers
Replicated with RAFT
Apache Ratis
If the size is < 5Gb
If everything is fine
Closed ContainersClosed Containers
Full containers
Immutable data!
Can be merged after
deleting data
49. 1 2 3 4 5 6 Datanodes
B
Block B is replaced to datanode: 1,2,3
1 1 1
50. 1 2 3 4 5 6 Datanodes
B
Block B is replaced to datanode: 1,2,3
ReplicaSet: Set of datanodes (3)
1 1 1
1 2 4
ReplicaSet (1,2,4) manages replicas of B block
1
83. Chance of data
loss (3 failure)
Source of replic.
(1 failure)
Random 2 group
6 group/
copysets
100% 3.5% 7%
all 2 4
SummarySummary
84. Summary (Copysets)Summary (Copysets)
Two main forces:
Number of distinct sets (copysets)
Number of source datanode to recover data
Random replication is not the most effective way
We can do better
Lower chance to loose data
More data to loose
85. Apache Hadoop OzoneApache Hadoop Ozone
S3 compatible Object store for Hadoop
On top of the Hadoop Distributed Storage layer
Advanced replication
Advanced block report handling (We small files)
Easy to use and run in containerized environments
88. Note: in-place upgradeNote: in-place upgrade
Upgrade path from existing HDFS
without (or minimal) Data Move
The math is the same
Grouping block in the same replicaset
Report them in groups (containers)
Better: Less replicasets with more blocks in each
Smart preprocessing/balancing may be required