This timely text/reference describes the development and implementation of large-scale distributed processing systems using open source tools and technologies such as Hadoop, Scalding and Spark.
Comprehensive in scope, the book presents state-of-the-art material on building high performance distributed computing systems, providing practical guidance and best practices as well as describing theoretical software frameworks.
Topics and features: describes the fundamentals of building scalable software systems for large-scale data processing in the new paradigm of high performance distributed computing; presents an overview of the Hadoop ecosystem, followed by step-by-step instruction on its installation, programming and execution; reviews the basics of Spark, including resilient distributed datasets, and examines Hadoop streaming and working with Scalding; provides detailed case studies on approaches to clustering, data classification and regression analysis; explains the process of creating a working recommender system using Scalding and Spark; supplies a complete list of supplementary source code and datasets at an associated website.
Fulfilling the need for both introductory material for undergraduate students of computer science and detailed discussions for software engineering professionals, this book will aid a broad audience to understand the esoteric aspects of practical high performance computing through its use of solved problems, research case studies and working source code.