CITESTE MAI MULT
Detalii
Descriere RO
The Web is getting faster, and the data it delivers is getting bigger. How can you handle everything efficiently? This book introduces Spark, an open source cluster computing system that makes data analytics fast to run and fast to write. You’ll learn how to run programs faster, using primitives for in-memory cluster computing. With Spark, your job can load data into memory and query it repeatedly much quicker than with disk-based systems like Hadoop MapReduce.
Written by the developers of Spark, this book will have you up and running in no time. You’ll learn how to express MapReduce jobs with just a few simple lines of Spark code, instead of spending extra time and effort working with Hadoop’s raw Java API.
Quickly dive into Spark capabilities such as collect, count, reduce, and save
Use one programming paradigm instead of mixing and matching tools such as Hive, Hadoop, Mahout, and S4/Storm
Learn how to run interactive, iterative, and incremental analyses
Integrate with Scala to manipulate distributed datasets like local collections
Tackle partitioning issues, data locality, default hash partitioning, user-defined partitioners, and custom serialization
Use other languages by means of pipe() to achieve the equivalent of Hadoop streaming
EdituraO'Reilly Media, Inc, USA
Dimensiuni180 x 233 x 14
Data Publicarii13/02/2015
Format
Necartonata
Numar pagini274
Aceasta este o carte in limba engleza. Descrierea cartii (tradusa din engleza cu Google Translate) este in limba romana din motive legale.
Web-ul devine mai rapid, iar datele pe care le furnizeaza sunt din ce in ce mai mari. Cum poti gestiona totul eficient?