85b50d4258
### What changes were proposed in this pull request? Cleanup all Zinc standalone server code, and realated coniguration. ### Why are the changes needed? ![image](https://user-images.githubusercontent.com/1736354/109154790-c1d3e580-77a9-11eb-8cde-835deed6e10e.png) - Zinc is the incremental compiler to speed up builds of compilation. - The scala-maven-plugin is the mave plugin, which is used by Spark, one of the function is to integrate the Zinc to enable the incremental compiler. - Since Spark v3.0.0 ([SPARK-28759](https://issues.apache.org/jira/browse/SPARK-28759)), the scala-maven-plugin is upgraded to v4.X, that means Zinc v0.3.13 standalone server is useless anymore. However, we still download, install, start the standalone Zinc server. we should remove all zinc standalone server code, and all related configuration. See more in [SPARK-34539](https://issues.apache.org/jira/projects/SPARK/issues/SPARK-34539) or the doc [Zinc standalone server is useless after scala-maven-plugin 4.x](https://docs.google.com/document/d/1u4kCHDx7KjVlHGerfmbcKSB0cZo6AD4cBdHSse-SBsM). ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? Run any mvn build: ./build/mvn -DskipTests clean package -pl core You could see the increamental compilation is still working, the stage of "scala-maven-plugin:4.3.0:compile (scala-compile-first)" with incremental compilation info, like: ``` [INFO] --- scala-maven-plugin:4.3.0:testCompile (scala-test-compile-first) spark-core_2.12 --- [INFO] Using incremental compilation using Mixed compile order [INFO] Compiler bridge file: /root/.sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.10__52.0-1.3.1_20191012T045515.jar [INFO] compiler plugin: BasicArtifact(com.github.ghik,silencer-plugin_2.12.10,1.6.0,null) [INFO] Compiling 303 Scala sources and 27 Java sources to /root/spark/core/target/scala-2.12/test-classes ... ``` Closes #31647 from Yikun/cleanup-zinc. Authored-by: Yikun Jiang <yikunkero@gmail.com> Signed-off-by: Sean Owen <srowen@gmail.com>
150 lines
4.4 KiB
Bash
Executable file
150 lines
4.4 KiB
Bash
Executable file
#!/usr/bin/env bash
|
|
|
|
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
#
|
|
# Creates a Spark release candidate. The script will update versions, tag the branch,
|
|
# build Spark binary packages and documentation, and upload maven artifacts to a staging
|
|
# repository. There is also a dry run mode where only local builds are performed, and
|
|
# nothing is uploaded to the ASF repos.
|
|
#
|
|
# Run with "-h" for options.
|
|
#
|
|
|
|
set -e
|
|
SELF=$(cd $(dirname $0) && pwd)
|
|
. "$SELF/release-util.sh"
|
|
|
|
function usage {
|
|
local NAME=$(basename $0)
|
|
cat <<EOF
|
|
Usage: $NAME [options]
|
|
|
|
This script runs the release scripts inside a docker image. The image is hardcoded to be called
|
|
"spark-rm" and will be re-generated (as needed) on every invocation of this script.
|
|
|
|
Options are:
|
|
|
|
-d [path] : required: working directory (output will be written to an "output" directory in
|
|
the working directory).
|
|
-n : dry run mode. Performs checks and local builds, but do not upload anything.
|
|
-t [tag] : tag for the spark-rm docker image to use for building (default: "latest").
|
|
-j [path] : path to local JDK installation to use for building. By default the script will
|
|
use openjdk8 installed in the docker image.
|
|
-s [step] : runs a single step of the process; valid steps are: tag, build, docs, publish
|
|
EOF
|
|
}
|
|
|
|
WORKDIR=
|
|
IMGTAG=latest
|
|
JAVA=
|
|
RELEASE_STEP=
|
|
while getopts ":d:hj:ns:t:" opt; do
|
|
case $opt in
|
|
d) WORKDIR="$OPTARG" ;;
|
|
n) DRY_RUN=1 ;;
|
|
t) IMGTAG="$OPTARG" ;;
|
|
j) JAVA="$OPTARG" ;;
|
|
s) RELEASE_STEP="$OPTARG" ;;
|
|
h) usage ;;
|
|
\?) error "Invalid option. Run with -h for help." ;;
|
|
esac
|
|
done
|
|
|
|
if [ -z "$WORKDIR" ] || [ ! -d "$WORKDIR" ]; then
|
|
error "Work directory (-d) must be defined and exist. Run with -h for help."
|
|
fi
|
|
|
|
if [ -d "$WORKDIR/output" ]; then
|
|
read -p "Output directory already exists. Overwrite and continue? [y/n] " ANSWER
|
|
if [ "$ANSWER" != "y" ]; then
|
|
error "Exiting."
|
|
fi
|
|
fi
|
|
|
|
cd "$WORKDIR"
|
|
rm -rf "$WORKDIR/output"
|
|
mkdir "$WORKDIR/output"
|
|
|
|
get_release_info
|
|
|
|
# Place all RM scripts and necessary data in a local directory that must be defined in the command
|
|
# line. This directory is mounted into the image.
|
|
for f in "$SELF"/*; do
|
|
if [ -f "$f" ]; then
|
|
cp "$f" "$WORKDIR"
|
|
fi
|
|
done
|
|
|
|
# Add the fallback version of Gemfile, Gemfile.lock and .bundle/config to the local directory.
|
|
cp "$SELF/../../docs/Gemfile" "$WORKDIR"
|
|
cp "$SELF/../../docs/Gemfile.lock" "$WORKDIR"
|
|
cp -r "$SELF/../../docs/.bundle" "$WORKDIR"
|
|
|
|
GPG_KEY_FILE="$WORKDIR/gpg.key"
|
|
fcreate_secure "$GPG_KEY_FILE"
|
|
$GPG --export-secret-key --armor --pinentry-mode loopback --passphrase "$GPG_PASSPHRASE" "$GPG_KEY" > "$GPG_KEY_FILE"
|
|
|
|
run_silent "Building spark-rm image with tag $IMGTAG..." "docker-build.log" \
|
|
docker build -t "spark-rm:$IMGTAG" --build-arg UID=$UID "$SELF/spark-rm"
|
|
|
|
# Write the release information to a file with environment variables to be used when running the
|
|
# image.
|
|
ENVFILE="$WORKDIR/env.list"
|
|
fcreate_secure "$ENVFILE"
|
|
|
|
function cleanup {
|
|
rm -f "$ENVFILE"
|
|
rm -f "$GPG_KEY_FILE"
|
|
}
|
|
|
|
trap cleanup EXIT
|
|
|
|
cat > $ENVFILE <<EOF
|
|
DRY_RUN=$DRY_RUN
|
|
SKIP_TAG=$SKIP_TAG
|
|
RUNNING_IN_DOCKER=1
|
|
GIT_BRANCH=$GIT_BRANCH
|
|
NEXT_VERSION=$NEXT_VERSION
|
|
RELEASE_VERSION=$RELEASE_VERSION
|
|
RELEASE_TAG=$RELEASE_TAG
|
|
GIT_REF=$GIT_REF
|
|
SPARK_PACKAGE_VERSION=$SPARK_PACKAGE_VERSION
|
|
ASF_USERNAME=$ASF_USERNAME
|
|
GIT_NAME=$GIT_NAME
|
|
GIT_EMAIL=$GIT_EMAIL
|
|
GPG_KEY=$GPG_KEY
|
|
ASF_PASSWORD=$ASF_PASSWORD
|
|
GPG_PASSPHRASE=$GPG_PASSPHRASE
|
|
RELEASE_STEP=$RELEASE_STEP
|
|
USER=$USER
|
|
EOF
|
|
|
|
JAVA_VOL=
|
|
if [ -n "$JAVA" ]; then
|
|
echo "JAVA_HOME=/opt/spark-java" >> $ENVFILE
|
|
JAVA_VOL="--volume $JAVA:/opt/spark-java"
|
|
fi
|
|
|
|
echo "Building $RELEASE_TAG; output will be at $WORKDIR/output"
|
|
docker run -ti \
|
|
--env-file "$ENVFILE" \
|
|
--volume "$WORKDIR:/opt/spark-rm" \
|
|
$JAVA_VOL \
|
|
"spark-rm:$IMGTAG"
|