spark-instrumented-optimizer/ec2
Dan Osipov b20171267d [SPARK-787] Add S3 configuration parameters to the EC2 deploy scripts
When deploying to AWS, there is additional configuration that is required to read S3 files. EMR creates it automatically, there is no reason that the Spark EC2 script shouldn't.

This PR requires a corresponding PR to the mesos/spark-ec2 to be merged, as it gets cloned in the process of setting up machines: https://github.com/mesos/spark-ec2/pull/58

Author: Dan Osipov <daniil.osipov@shazam.com>

Closes #1120 from danosipov/s3_credentials and squashes the following commits:

758da8b [Dan Osipov] Modify documentation to include the new parameter
71fab14 [Dan Osipov] Use a parameter --copy-aws-credentials to enable S3 credential deployment
7e0da26 [Dan Osipov] Get AWS credentials out of boto connection instance
39bdf30 [Dan Osipov] Add S3 configuration parameters to the EC2 deploy scripts
2014-09-16 13:40:16 -07:00
..
deploy.generic/root/spark-ec2 [SPARK-787] Add S3 configuration parameters to the EC2 deploy scripts 2014-09-16 13:40:16 -07:00
third_party Added script for launching Spark on EC2 from Mesos, to make it easier 2012-06-10 23:06:15 -07:00
README Remove remaining references to incubation 2014-03-02 01:00:16 -08:00
spark-ec2 SPARK-2241: quote command line args in ec2 script 2014-06-22 20:23:49 -07:00
spark_ec2.py [SPARK-787] Add S3 configuration parameters to the EC2 deploy scripts 2014-09-16 13:40:16 -07:00

This folder contains a script, spark-ec2, for launching Spark clusters on
Amazon EC2. Usage instructions are available online at:

http://spark.apache.org/docs/latest/ec2-scripts.html