1450b5e095
### What changes were proposed in this pull request? Fix typo for docs, log messages and comments ### Why are the changes needed? typo fix to increase readability ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? manual test has been performed to test the updated Closes #29443 from brandonJY/spell-fix-doc. Authored-by: Brandon Jiang <Brandon.jiang.a@outlook.com> Signed-off-by: Takeshi Yamamuro <yamamuro@apache.org>
58 lines
2 KiB
Bash
Executable file
58 lines
2 KiB
Bash
Executable file
#!/usr/bin/env bash
|
|
|
|
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
# A shell script to decommission all workers on a single worker
|
|
#
|
|
# Environment variables
|
|
#
|
|
# SPARK_WORKER_INSTANCES The number of worker instances that should be
|
|
# running on this worker machine. Default is 1.
|
|
|
|
# Usage: decommission-worker.sh [--block-until-exit]
|
|
# Decommissions all workers on this worker machine.
|
|
|
|
set -ex
|
|
|
|
if [ -z "${SPARK_HOME}" ]; then
|
|
export SPARK_HOME="$(cd "`dirname "$0"`"/..; pwd)"
|
|
fi
|
|
|
|
. "${SPARK_HOME}/sbin/spark-config.sh"
|
|
|
|
. "${SPARK_HOME}/bin/load-spark-env.sh"
|
|
|
|
if [ "$SPARK_WORKER_INSTANCES" = "" ]; then
|
|
"${SPARK_HOME}/sbin"/spark-daemon.sh decommission org.apache.spark.deploy.worker.Worker 1
|
|
else
|
|
for ((i=0; i<$SPARK_WORKER_INSTANCES; i++)); do
|
|
"${SPARK_HOME}/sbin"/spark-daemon.sh decommission org.apache.spark.deploy.worker.Worker $(( $i + 1 ))
|
|
done
|
|
fi
|
|
|
|
# Check if --block-until-exit is set.
|
|
# This is done for systems which block on the decommissioning script and on exit
|
|
# shut down the entire system (e.g. K8s).
|
|
if [ "$1" == "--block-until-exit" ]; then
|
|
shift
|
|
# For now we only block on the 0th instance if there multiple instances.
|
|
instance=$1
|
|
pid="$SPARK_PID_DIR/spark-$SPARK_IDENT_STRING-$command-$instance.pid"
|
|
wait $pid
|
|
fi
|