<--- Back to Details
First PageDocument Content
Computing / Hadoop / Apache Software Foundation / Parallel computing / Cluster computing / Java platform / Apache Spark / MapReduce / Data-intensive computing / Apache Hadoop / Apache Hive / Scala
Date: 2016-08-21 15:09:53
Computing
Hadoop
Apache Software Foundation
Parallel computing
Cluster computing
Java platform
Apache Spark
MapReduce
Data-intensive computing
Apache Hadoop
Apache Hive
Scala

Spark: Cluster Computing with Working Sets Matei Zaharia, Mosharaf Chowdhury, Michael J. Franklin, Scott Shenker, Ion Stoica University of California, Berkeley Abstract MapReduce/Dryad job, each job must reload the data

Add to Reading List

Source URL: people.csail.mit.edu

Download Document from Source Website

File Size: 205,21 KB

Share Document on Facebook

Similar Documents

Accelerating Big Data Processing with Hadoop, Spark and Memcached Talk at HPC Advisory Council Switzerland Conference (Mar ‘15) by Dhabaleswar K. (DK) Panda The Ohio State University

Accelerating Big Data Processing with Hadoop, Spark and Memcached Talk at HPC Advisory Council Switzerland Conference (Mar ‘15) by Dhabaleswar K. (DK) Panda The Ohio State University

DocID: 1xVBp - View Document

Big Data Hadoop Certification Training About Intellipaat Intellipaat is a fast-growing professional training provider that is offering training in over 150 most sought-after tools and technologies. We have a learner base

Big Data Hadoop Certification Training About Intellipaat Intellipaat is a fast-growing professional training provider that is offering training in over 150 most sought-after tools and technologies. We have a learner base

DocID: 1xVAd - View Document

PDF Document

DocID: 1xPx3 - View Document

PDF Document

DocID: 1xGCT - View Document

PDF Document

DocID: 1xvV8 - View Document