007a733434
Various spark scripts load spark-env.sh. This can cause growth of any variables that may be appended to (SPARK_CLASSPATH, SPARK_REPL_OPTS) and it makes the precedence order for options specified in spark-env.sh less clear. One use-case for the latter is that we want to set options from the command-line of spark-shell, but these options will be overridden by subsequent loading of spark-env.sh. If we were to load the spark-env.sh first and then set our command-line options, we could guarantee correct precedence order. Note that we use SPARK_CONF_DIR if available to support the sbin/ scripts, which always set this variable from sbin/spark-config.sh. Otherwise, we default to the ../conf/ as usual. Author: Aaron Davidson <aaron@databricks.com> Closes #184 from aarondav/idem and squashes the following commits: e291f91 [Aaron Davidson] Use "private" variables in load-spark-env.sh 8da8360 [Aaron Davidson] Add .sh extension to load-spark-env.sh 93a2471 [Aaron Davidson] SPARK-1286: Make usage of spark-env.sh idempotent
70 lines
2.2 KiB
Bash
Executable file
70 lines
2.2 KiB
Bash
Executable file
#!/usr/bin/env bash
|
|
|
|
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
sbin=`dirname "$0"`
|
|
sbin=`cd "$sbin"; pwd`
|
|
|
|
|
|
START_TACHYON=false
|
|
|
|
while (( "$#" )); do
|
|
case $1 in
|
|
--with-tachyon)
|
|
if [ ! -e "$sbin"/../tachyon/bin/tachyon ]; then
|
|
echo "Error: --with-tachyon specified, but tachyon not found."
|
|
exit -1
|
|
fi
|
|
START_TACHYON=true
|
|
;;
|
|
esac
|
|
shift
|
|
done
|
|
|
|
. "$sbin/spark-config.sh"
|
|
|
|
. "$SPARK_PREFIX/bin/load-spark-env.sh"
|
|
|
|
# Find the port number for the master
|
|
if [ "$SPARK_MASTER_PORT" = "" ]; then
|
|
SPARK_MASTER_PORT=7077
|
|
fi
|
|
|
|
if [ "$SPARK_MASTER_IP" = "" ]; then
|
|
SPARK_MASTER_IP=`hostname`
|
|
fi
|
|
|
|
if [ "$START_TACHYON" == "true" ]; then
|
|
"$sbin/slaves.sh" cd "$SPARK_HOME" \; "$sbin"/../tachyon/bin/tachyon bootstrap-conf $SPARK_MASTER_IP
|
|
|
|
# set -t so we can call sudo
|
|
SPARK_SSH_OPTS="-o StrictHostKeyChecking=no -t" "$sbin/slaves.sh" cd "$SPARK_HOME" \; "$sbin/../tachyon/bin/tachyon-start.sh" worker SudoMount \; sleep 1
|
|
fi
|
|
|
|
# Launch the slaves
|
|
if [ "$SPARK_WORKER_INSTANCES" = "" ]; then
|
|
exec "$sbin/slaves.sh" cd "$SPARK_HOME" \; "$sbin/start-slave.sh" 1 spark://$SPARK_MASTER_IP:$SPARK_MASTER_PORT
|
|
else
|
|
if [ "$SPARK_WORKER_WEBUI_PORT" = "" ]; then
|
|
SPARK_WORKER_WEBUI_PORT=8081
|
|
fi
|
|
for ((i=0; i<$SPARK_WORKER_INSTANCES; i++)); do
|
|
"$sbin/slaves.sh" cd "$SPARK_HOME" \; "$sbin/start-slave.sh" $(( $i + 1 )) spark://$SPARK_MASTER_IP:$SPARK_MASTER_PORT --webui-port $(( $SPARK_WORKER_WEBUI_PORT + $i ))
|
|
done
|
|
fi
|