Apache Spark - A unified analytics engine for large-scale data processing
Go to file
Mosharaf Chowdhury 4f0b7eb02d SplitStream was not working in EC2.
We have turned OFF SSB for now.
2010-05-16 13:00:47 -07:00
src SplitStream was not working in EC2. 2010-05-16 13:00:47 -07:00
third_party SplitStream implementation in progress. 2010-04-19 00:14:53 -07:00
.gitignore Initial commit 2010-03-29 16:17:55 -07:00
alltests Imported Mosharaf's multi-tracker branch 2010-04-03 23:50:04 -07:00
lr_data.txt Initial commit 2010-03-29 16:17:55 -07:00
Makefile SplitStream implementation in progress. 2010-04-19 00:14:53 -07:00
README Initial commit 2010-03-29 16:17:55 -07:00
run SplitStream integration in progress. 2010-04-20 02:08:48 -07:00
spark-executor Initial commit 2010-03-29 16:17:55 -07:00
spark-shell Initial commit 2010-03-29 16:17:55 -07:00

Spark requires Scala 2.7.7. It will currently not work with 2.8, or with
earlier versions of the 2.7 branch.

To build and run Spark, you will need to have Scala's bin in your $PATH,
or you will need to set the SCALA_HOME environment variable to point
to where you've installed Scala. Scala must be accessible through one
of these methods on Nexus slave nodes as well as on the master.

To build Spark and the example programs, run make.

To run one of the examples, use ./run <class> <params>. For example,
./run SparkLR will run the Logistic Regression example. Each of the
example programs prints usage help if no params are given.

Tip: If you are building Spark and examples repeatedly, export USE_FSC=1
to have the Makefile use the fsc compiler daemon instead of scalac.