Get ready to unlock the power of your data. With the fourth edition of this comprehensive guide, you’ll learn how to build and maintain reliable, scalable, distributed systems with Apache Hadoop. This book is ideal for programmers looking to analyze datasets of any size, and for administrators who want to set up and run Hadoop clusters.
Using Hadoop 2 exclusively, author Tom White presents new chapters on YARN and several Hadoop-related projects such as Parquet, Flume, Crunch, and Spark. You’ll learn about recent changes to Hadoop, and explore new case studies on Hadoop’s role in healthcare systems and genomics data processing.
Learn fundamental components such as MapReduce, HDFS, and YARN.
Explore MapReduce in depth, including steps for developing applications with it.
Set up and maintain a Hadoop cluster running HDFS and MapReduce on YARN.
Learn two data formats: Avro for data serialization and Parquet for nested data.
Use data ingestion tools such as Flume (for streaming data) and Sqoop (for bulk data transfer).
Understand how high-level data processing tools like Pig, Hive, Crunch, and Spark work with Hadoop.
Learn the HBase distributed database and the ZooKeeper distributed configuration service.