Gehören Sie zu den Ersten, denen das gefällt!
This presentation about Hadoop YARN will help you understand the Hadoop 1.0 and Hadoop 2.0, limitations of Hadoop 1.0, need for YARN, what is YARN, workloads running on YARN, YARN components, YARN architecture and you will also go through a demo on YARN. YARN is the cluster resource management layer of the Apache Hadoop Ecosystem, which schedules jobs and assigns resources. Hadoop 1.0 is designed to run MapReduce jobs only and had issues in scalability, resource utilization, etc. whereas YARN solved those issues and users could work on multiple processing models. Now let us get started and learn YARN in detail.
Below topics are explained in this Hadoop YARN presentation:
1. Hadoop 1.0 (MapReduce 1)
2. Limitations of Hadoop 1.0 (MapReduce 1)
3. Need for YARN
4. What is YARN
5. Workloads running on YARN
6. YARN components
7. YARN architecture
8. Demo on YARN
What is this Big Data Hadoop training course about?
The Big Data Hadoop and Spark developer course have been designed to impart an in-depth knowledge of Big Data processing using Hadoop and Spark. The course is packed with real-life projects and case studies to be executed in the CloudLab.
What are the course objectives?
This course will enable you to:
1. Understand the different components of the Hadoop ecosystem such as Hadoop 2.7, Yarn, MapReduce, Pig, Hive, Impala, HBase, Sqoop, Flume, and Apache Spark
2. Understand Hadoop Distributed File System (HDFS) and YARN as well as their architecture, and learn how to work with them for storage and resource management
3. Understand MapReduce and its characteristics, and assimilate some advanced MapReduce concepts
4. Get an overview of Sqoop and Flume and describe how to ingest data using them
5. Create database and tables in Hive and Impala, understand HBase, and use Hive and Impala for partitioning
6. Understand different types of file formats, Avro Schema, using Arvo with Hive, and Sqoop and Schema evolution
7. Understand Flume, Flume architecture, sources, flume sinks, channels, and flume configurations
8. Understand HBase, its architecture, data storage, and working with HBase. You will also understand the difference between HBase and RDBMS
9. Gain a working knowledge of Pig and its components
10. Do functional programming in Spark
11. Understand resilient distribution datasets (RDD) in detail
12. Implement and build Spark applications
13. Gain an in-depth understanding of parallel processing in Spark and Spark RDD optimization techniques
14. Understand the common use-cases of Spark and the various interactive algorithms
15. Learn Spark SQL, creating, transforming, and querying Data frames
Learn more at https://www.simplilearn.com/big-data-and-analytics/big-data-and-hadoop-training