spark-instrumented-optimizer/ec2
Daniel Darabos 44d3a6a752 [SPARK-3342] Add SSDs to block device mapping
On `m3.2xlarge` instances the 2x80GB SSDs are inaccessible if not added to the block device mapping when the instance is created. They work when added with this patch. I have not tested this with other instance types, and I do not know much about this script and EC2 deployment in general. Maybe this code needs to depend on the instance type.

The requirement for this mapping is described in the AWS docs at:
http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/InstanceStorage.html#InstanceStore_UsageScenarios

"For M3 instances, you must specify instance store volumes in the block
device mapping for the instance. When you launch an M3 instance, we
ignore any instance store volumes specified in the block device mapping
for the AMI."

Author: Daniel Darabos <darabos.daniel@gmail.com>

Closes #2081 from darabos/patch-1 and squashes the following commits:

1ceb2c8 [Daniel Darabos] Use %d string interpolation instead of {}.
a1854d7 [Daniel Darabos] Only specify ephemeral device mapping for M3.
e0d9e37 [Daniel Darabos] Create ephemeral device mapping based on get_num_disks().
6b116a6 [Daniel Darabos] Add SSDs to block device mapping
2014-09-01 22:16:42 -07:00
..
deploy.generic/root/spark-ec2 EC2 configurable workers 2014-05-03 16:52:19 -07:00
third_party Added script for launching Spark on EC2 from Mesos, to make it easier 2012-06-10 23:06:15 -07:00
README Remove remaining references to incubation 2014-03-02 01:00:16 -08:00
spark-ec2 SPARK-2241: quote command line args in ec2 script 2014-06-22 20:23:49 -07:00
spark_ec2.py [SPARK-3342] Add SSDs to block device mapping 2014-09-01 22:16:42 -07:00

This folder contains a script, spark-ec2, for launching Spark clusters on
Amazon EC2. Usage instructions are available online at:

http://spark.apache.org/docs/latest/ec2-scripts.html