Get Free Shipping on orders over $79
Learn Apache Spark Full - Hoang Tran

Learn Apache Spark Full

By: Hoang Tran

eBook | 28 February 2020

At a Glance

eBook


$5.05

or 4 interest-free payments of $1.26 with

Instant Digital Delivery to your Kobo Reader App

Industries are using Hadoop extensively to analyze their data sets. The reason is that Hadoop framework is based on a simple programming model (MapReduce) and it enables a computing solution that is scalable, flexible, fault-tolerant and cost effective. Here, the main concern is to maintain speed in processing large datasets in terms of waiting time between queries and waiting time to run the program.

Spark was introduced by Apache Software Foundation for speeding up the Hadoop computational computing software process.

As against a common belief, Spark is not a modified version of Hadoop and is not, really, dependent on Hadoop because it has its own cluster management. Hadoop is just one of the ways to implement Spark.

Spark uses Hadoop in two ways - one is storage and second is processing. Since Spark has its own cluster management computation, it uses Hadoop for storage purpose only.

on

More in Computer Science

Amazon.com : Get Big Fast - Robert Spector

eBOOK

AI-Powered Search - Trey Grainger

eBOOK

Tissue Proteomics : Methods and Protocols - Taufika Islam Williams

eBOOK

RRP $369.00

$332.99

10%
OFF