Apache Hadoop is a popular framework for enabling applications to collect data from various locations and in various formats. Businesses use it to facilitate the execution of distributed processes against large amounts of data. It achieves this by allowing the clustering of multiple computers to analyze massive datasets quickly. During this course, you will be introduced to the basics of Apache Hadoop. You will explore its architecture and core components, including MapReduce and the Hadoop Distributed File System (HDFS). You will learn how to add and remove nodes from Hadoop clusters, how to check available disk space on each node, and how to modify configuration parameters. Plus, you will learn about other Apache projects that are part of the Hadoop ecosystem, including Pig, Hive, HBase, ZooKeeper, Oozie, Sqoop, and Flume. For individuals keen to build core skills for the big data domain, therefore, Hadoop 101 is an excellent place to start for this leg of your data science journey.