53cd50c069
This commit makes Spark invocation saner by using an assembly JAR to find all of Spark's dependencies instead of adding all the JARs in lib_managed. It also packages the examples into an assembly and uses that as SPARK_EXAMPLES_JAR. Finally, it replaces the old "run" script with two better-named scripts: "run-examples" for examples, and "spark-class" for Spark internal classes (e.g. REPL, master, etc). This is also designed to minimize the confusion people have in trying to use "run" to run their own classes; it's not meant to do that, but now at least if they look at it, they can modify run-examples to do a decent job for them. As part of this, Bagel's examples are also now properly moved to the examples package instead of bagel.
158 lines
3.9 KiB
Bash
Executable file
158 lines
3.9 KiB
Bash
Executable file
#!/usr/bin/env bash
|
|
|
|
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
# Runs a Spark command as a daemon.
|
|
#
|
|
# Environment Variables
|
|
#
|
|
# SPARK_CONF_DIR Alternate conf dir. Default is ${SPARK_PREFIX}/conf.
|
|
# SPARK_LOG_DIR Where log files are stored. PWD by default.
|
|
# SPARK_MASTER host:path where spark code should be rsync'd from
|
|
# SPARK_PID_DIR The pid files are stored. /tmp by default.
|
|
# SPARK_IDENT_STRING A string representing this instance of spark. $USER by default
|
|
# SPARK_NICENESS The scheduling priority for daemons. Defaults to 0.
|
|
##
|
|
|
|
usage="Usage: spark-daemon.sh [--config <conf-dir>] [--hosts hostlistfile] (start|stop) <spark-command> <spark-instance-number> <args...>"
|
|
|
|
# if no args specified, show usage
|
|
if [ $# -le 1 ]; then
|
|
echo $usage
|
|
exit 1
|
|
fi
|
|
|
|
bin=`dirname "$0"`
|
|
bin=`cd "$bin"; pwd`
|
|
|
|
. "$bin/spark-config.sh"
|
|
|
|
# get arguments
|
|
startStop=$1
|
|
shift
|
|
command=$1
|
|
shift
|
|
instance=$1
|
|
shift
|
|
|
|
spark_rotate_log ()
|
|
{
|
|
log=$1;
|
|
num=5;
|
|
if [ -n "$2" ]; then
|
|
num=$2
|
|
fi
|
|
if [ -f "$log" ]; then # rotate logs
|
|
while [ $num -gt 1 ]; do
|
|
prev=`expr $num - 1`
|
|
[ -f "$log.$prev" ] && mv "$log.$prev" "$log.$num"
|
|
num=$prev
|
|
done
|
|
mv "$log" "$log.$num";
|
|
fi
|
|
}
|
|
|
|
if [ -f "${SPARK_CONF_DIR}/spark-env.sh" ]; then
|
|
. "${SPARK_CONF_DIR}/spark-env.sh"
|
|
fi
|
|
|
|
if [ "$SPARK_IDENT_STRING" = "" ]; then
|
|
export SPARK_IDENT_STRING="$USER"
|
|
fi
|
|
|
|
export SPARK_PRINT_LAUNCH_COMMAND="1"
|
|
|
|
# get log directory
|
|
if [ "$SPARK_LOG_DIR" = "" ]; then
|
|
export SPARK_LOG_DIR="$SPARK_HOME/logs"
|
|
fi
|
|
mkdir -p "$SPARK_LOG_DIR"
|
|
touch $SPARK_LOG_DIR/.spark_test > /dev/null 2>&1
|
|
TEST_LOG_DIR=$?
|
|
if [ "${TEST_LOG_DIR}" = "0" ]; then
|
|
rm -f $SPARK_LOG_DIR/.spark_test
|
|
else
|
|
chown $SPARK_IDENT_STRING $SPARK_LOG_DIR
|
|
fi
|
|
|
|
if [ "$SPARK_PID_DIR" = "" ]; then
|
|
SPARK_PID_DIR=/tmp
|
|
fi
|
|
|
|
# some variables
|
|
export SPARK_LOGFILE=spark-$SPARK_IDENT_STRING-$command-$instance-$HOSTNAME.log
|
|
export SPARK_ROOT_LOGGER="INFO,DRFA"
|
|
log=$SPARK_LOG_DIR/spark-$SPARK_IDENT_STRING-$command-$instance-$HOSTNAME.out
|
|
pid=$SPARK_PID_DIR/spark-$SPARK_IDENT_STRING-$command-$instance.pid
|
|
|
|
# Set default scheduling priority
|
|
if [ "$SPARK_NICENESS" = "" ]; then
|
|
export SPARK_NICENESS=0
|
|
fi
|
|
|
|
|
|
case $startStop in
|
|
|
|
(start)
|
|
|
|
mkdir -p "$SPARK_PID_DIR"
|
|
|
|
if [ -f $pid ]; then
|
|
if kill -0 `cat $pid` > /dev/null 2>&1; then
|
|
echo $command running as process `cat $pid`. Stop it first.
|
|
exit 1
|
|
fi
|
|
fi
|
|
|
|
if [ "$SPARK_MASTER" != "" ]; then
|
|
echo rsync from $SPARK_MASTER
|
|
rsync -a -e ssh --delete --exclude=.svn --exclude='logs/*' --exclude='contrib/hod/logs/*' $SPARK_MASTER/ "$SPARK_HOME"
|
|
fi
|
|
|
|
spark_rotate_log $log
|
|
echo starting $command, logging to $log
|
|
echo "Spark Daemon: $command" > $log
|
|
cd "$SPARK_PREFIX"
|
|
nohup nice -n $SPARK_NICENESS "$SPARK_PREFIX"/spark-class $command "$@" >> "$log" 2>&1 < /dev/null &
|
|
echo $! > $pid
|
|
sleep 1; head "$log"
|
|
;;
|
|
|
|
(stop)
|
|
|
|
if [ -f $pid ]; then
|
|
if kill -0 `cat $pid` > /dev/null 2>&1; then
|
|
echo stopping $command
|
|
kill `cat $pid`
|
|
else
|
|
echo no $command to stop
|
|
fi
|
|
else
|
|
echo no $command to stop
|
|
fi
|
|
;;
|
|
|
|
(*)
|
|
echo $usage
|
|
exit 1
|
|
;;
|
|
|
|
esac
|
|
|
|
|