Apache Hadoop is one of the hottest technologies that paves the ground for analyzing big data. Learn more about what Hadoop is and its components, such as MapReduce and HDFS. Come on this journey to play with large data sets and see Hadoop’s method of distributed processing.
Are you interested in moving beyond the elephant in the room and understanding Hadoop as a foundational tool set in your future? Then select this learning path to gain exposure to the tools used in Big Data, Hadoop's core components and supporting open source projects.
About this learning path
Hadoop is a name that represents two items, one a child’s toy and the other an open source framework for distributed storage and processing of big data. In both contexts, interaction with Hadoop is foundational in personal growth and development. This learning path covers content that is critical to your success in this realm. It takes you on a journey that explains the Hadoop conceptual design, then it looks how to use the application and then manipulate data without the use of complex coding.
Come along and start your journey to receiving the following badges: Hadoop Foundations – Level 1, Hadoop Programming – Level 1, Hadoop Administration – Level 1, Hadoop Data Access – Level 1, and Hadoop Foundations – Level 2!
String together your understanding of Yet Another Resource Negotiator (YARN) by gaining exposure to MapReduce1, the tool-sets that start the processing of Big Data.
Open the door to move data into Hadoop to get the program working for you. This course’s emphasis on Sqoop and Flume is on purpose to get you where you need to be obstacle free.
Writing MapReduce programs to analyze Big Data can get complex. In this Accessing Hadoop Data Using Hive course, you will get a solid foundation on using Apache Hive, a tool that can help make querying your data much easier. You will learn how to query, summarize, and analyze large data sets stored in Hadoop compatible file systems.