Are you interested in moving beyond the elephant in the room and understanding Hadoop as a foundational tool set in your future? Then select this learning path to gain exposure to the tools used in Big Data, Hadoop's core components and supporting open source projects.
Hadoop is a name that represents two items, one a child’s toy and the other an open source framework for distributed storage and processing of big data. In both contexts, interaction with Hadoop is foundational in personal growth and development. This learning path covers content that is critical to your success in this realm. It takes you on a journey that explains the Hadoop conceptual design, then it looks how to use the application and then manipulate data without the use of complex coding.
Apache Hadoop is one of the hottest technologies that paves the ground for analyzing big data. Learn more about what Hadoop is and its components, such as MapReduce and HDFS. Come on this journey to play with large data sets and see Hadoop’s method of distributed processing.
String together your understanding of Yet Another Resource Negotiator (YARN) by gaining exposure to MapReduce1, the tool-sets that start the processing of Big Data.
MapReduce and YARN