Back to Results
First PageMeta Content
Computing / Hadoop / Apache Software Foundation / Parallel computing / Cluster computing / Java platform / Apache Spark / MapReduce / Data-intensive computing / Apache Hadoop / Apache Hive / Scala


Spark: Cluster Computing with Working Sets Matei Zaharia, Mosharaf Chowdhury, Michael J. Franklin, Scott Shenker, Ion Stoica University of California, Berkeley Abstract MapReduce/Dryad job, each job must reload the data
Add to Reading List

Document Date: 2016-08-21 15:09:53


Open Document

File Size: 205,21 KB

Share Result on Facebook