Skip to content
Snippets Groups Projects
Commit 5b7ee173 authored by Matei Zaharia's avatar Matei Zaharia
Browse files

Update EC2 scripts for Spark 0.6

parent b4067cba
No related branches found
No related tags found
No related merge requests found
...@@ -6,3 +6,4 @@ export MESOS_SLAVES="{{slave_list}}" ...@@ -6,3 +6,4 @@ export MESOS_SLAVES="{{slave_list}}"
export MESOS_ZOO_LIST="{{zoo_list}}" export MESOS_ZOO_LIST="{{zoo_list}}"
export MESOS_HDFS_DATA_DIRS="{{hdfs_data_dirs}}" export MESOS_HDFS_DATA_DIRS="{{hdfs_data_dirs}}"
export MESOS_MAPRED_LOCAL_DIRS="{{mapred_local_dirs}}" export MESOS_MAPRED_LOCAL_DIRS="{{mapred_local_dirs}}"
export MESOS_SPARK_LOCAL_DIRS="{{spark_local_dirs}}"
...@@ -35,8 +35,7 @@ from boto.ec2.blockdevicemapping import BlockDeviceMapping, EBSBlockDeviceType ...@@ -35,8 +35,7 @@ from boto.ec2.blockdevicemapping import BlockDeviceMapping, EBSBlockDeviceType
# A static URL from which to figure out the latest Mesos EC2 AMI # A static URL from which to figure out the latest Mesos EC2 AMI
LATEST_AMI_URL = "https://s3.amazonaws.com/mesos-images/ids/latest-spark-0.5" LATEST_AMI_URL = "https://s3.amazonaws.com/mesos-images/ids/latest-spark-0.6"
LATEST_STANDALONE_AMI_URL = "https://s3.amazonaws.com/spark-standalone-amis/latest-spark"
# Configure and parse our command-line arguments # Configure and parse our command-line arguments
...@@ -65,8 +64,7 @@ def parse_args(): ...@@ -65,8 +64,7 @@ def parse_args():
help="Availability zone to launch instances in") help="Availability zone to launch instances in")
parser.add_option("-a", "--ami", default="latest", parser.add_option("-a", "--ami", default="latest",
help="Amazon Machine Image ID to use, or 'latest' to use latest " + help="Amazon Machine Image ID to use, or 'latest' to use latest " +
"availabe mesos AMI, 'standalone' for the latest available " + "available AMI (default: latest)")
"standalone AMI (default: latest)")
parser.add_option("-D", metavar="[ADDRESS:]PORT", dest="proxy_port", parser.add_option("-D", metavar="[ADDRESS:]PORT", dest="proxy_port",
help="Use SSH dynamic port forwarding to create a SOCKS proxy at " + help="Use SSH dynamic port forwarding to create a SOCKS proxy at " +
"the given local address (for use with login)") "the given local address (for use with login)")
...@@ -193,19 +191,14 @@ def launch_cluster(conn, opts, cluster_name): ...@@ -193,19 +191,14 @@ def launch_cluster(conn, opts, cluster_name):
"group %s, %s or %s" % (master_group.name, slave_group.name, zoo_group.name)) "group %s, %s or %s" % (master_group.name, slave_group.name, zoo_group.name))
sys.exit(1) sys.exit(1)
if opts.ami in ["latest", "standalone"]: # Figure out the latest AMI from our static URL
if opts.ami == "latest":
# Figure out the latest AMI from our static URL
if opts.ami == "latest":
url = LATEST_AMI_URL
elif opts.ami == "standalone":
url = LATEST_STANDALONE_AMI_URL
try: try:
opts.ami = urllib2.urlopen(url).read().strip() opts.ami = urllib2.urlopen(LATEST_AMI_URL).read().strip()
print "Latest Spark AMI: " + opts.ami print "Latest Spark AMI: " + opts.ami
except: except:
print >> stderr, "Could not read " + url print >> stderr, "Could not read " + LATEST_AMI_URL
sys.exit(1)
print "Launching instances..." print "Launching instances..."
...@@ -403,10 +396,12 @@ def deploy_files(conn, root_dir, opts, master_nodes, slave_nodes, zoo_nodes): ...@@ -403,10 +396,12 @@ def deploy_files(conn, root_dir, opts, master_nodes, slave_nodes, zoo_nodes):
num_disks = get_num_disks(opts.instance_type) num_disks = get_num_disks(opts.instance_type)
hdfs_data_dirs = "/mnt/ephemeral-hdfs/data" hdfs_data_dirs = "/mnt/ephemeral-hdfs/data"
mapred_local_dirs = "/mnt/hadoop/mrlocal" mapred_local_dirs = "/mnt/hadoop/mrlocal"
spark_local_dirs = "/mnt/spark"
if num_disks > 1: if num_disks > 1:
for i in range(2, num_disks + 1): for i in range(2, num_disks + 1):
hdfs_data_dirs += ",/mnt%d/ephemeral-hdfs/data" % i hdfs_data_dirs += ",/mnt%d/ephemeral-hdfs/data" % i
mapred_local_dirs += ",/mnt%d/hadoop/mrlocal" % i mapred_local_dirs += ",/mnt%d/hadoop/mrlocal" % i
spark_local_dirs += ",/mnt%d/spark" % i
if zoo_nodes != []: if zoo_nodes != []:
zoo_list = '\n'.join([i.public_dns_name for i in zoo_nodes]) zoo_list = '\n'.join([i.public_dns_name for i in zoo_nodes])
...@@ -426,7 +421,8 @@ def deploy_files(conn, root_dir, opts, master_nodes, slave_nodes, zoo_nodes): ...@@ -426,7 +421,8 @@ def deploy_files(conn, root_dir, opts, master_nodes, slave_nodes, zoo_nodes):
"zoo_list": zoo_list, "zoo_list": zoo_list,
"cluster_url": cluster_url, "cluster_url": cluster_url,
"hdfs_data_dirs": hdfs_data_dirs, "hdfs_data_dirs": hdfs_data_dirs,
"mapred_local_dirs": mapred_local_dirs "mapred_local_dirs": mapred_local_dirs,
"spark_local_dirs": spark_local_dirs
} }
# Create a temp directory in which we will place all the files to be # Create a temp directory in which we will place all the files to be
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment