Gain practical, hands-on experience installing and running Hadoop and Spark on your own desktop or laptop, and progress to managing real-world cluster deployments. Through engaging lessons and interactive examples, you’ll master essential concepts such as HDFS, MapReduce, PySpark, HiveQL, and data ingestion tools, while also learning to leverage user-friendly interfaces like Ambari and Zeppelin to streamline analytics workflows and cluster administration. By the end of this course, you’ll possess the foundational skills and confidence to begin your journey in big data analytics and explore the vast Hadoop ecosystem.
Applied Learning Project
Run Apache Pig, Hive, Flume, Sqoop, Oozie, Spark applications and write basic MapReduce/Spark programs. The steps for easily installing a working Hadoop/Spark system on a desktop/laptop and on a local stand-alone cluster using the powerful Ambari GUI are also included. All software used is open source and freely available.