Make deprecation warning less severe
Just a small change. I think it's good not to scare people who are using the old options. Author: Patrick Wendell <pwendell@gmail.com> Closes #810 from pwendell/warnings and squashes the following commits: cb8a311 [Patrick Wendell] Make deprecation warning less severe
This commit is contained in:
parent
cf6cbe9f76
commit
442808a748
|
@ -238,10 +238,10 @@ class SparkConf(loadDefaults: Boolean) extends Cloneable with Logging {
|
||||||
|
|
||||||
// Check for legacy configs
|
// Check for legacy configs
|
||||||
sys.env.get("SPARK_JAVA_OPTS").foreach { value =>
|
sys.env.get("SPARK_JAVA_OPTS").foreach { value =>
|
||||||
val error =
|
val warning =
|
||||||
s"""
|
s"""
|
||||||
|SPARK_JAVA_OPTS was detected (set to '$value').
|
|SPARK_JAVA_OPTS was detected (set to '$value').
|
||||||
|This has undefined behavior when running on a cluster and is deprecated in Spark 1.0+.
|
|This is deprecated in Spark 1.0+.
|
||||||
|
|
|
|
||||||
|Please instead use:
|
|Please instead use:
|
||||||
| - ./spark-submit with conf/spark-defaults.conf to set defaults for an application
|
| - ./spark-submit with conf/spark-defaults.conf to set defaults for an application
|
||||||
|
@ -249,7 +249,7 @@ class SparkConf(loadDefaults: Boolean) extends Cloneable with Logging {
|
||||||
| - spark.executor.extraJavaOptions to set -X options for executors
|
| - spark.executor.extraJavaOptions to set -X options for executors
|
||||||
| - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master or worker)
|
| - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master or worker)
|
||||||
""".stripMargin
|
""".stripMargin
|
||||||
logError(error)
|
logWarning(warning)
|
||||||
|
|
||||||
for (key <- Seq(executorOptsKey, driverOptsKey)) {
|
for (key <- Seq(executorOptsKey, driverOptsKey)) {
|
||||||
if (getOption(key).isDefined) {
|
if (getOption(key).isDefined) {
|
||||||
|
@ -262,16 +262,16 @@ class SparkConf(loadDefaults: Boolean) extends Cloneable with Logging {
|
||||||
}
|
}
|
||||||
|
|
||||||
sys.env.get("SPARK_CLASSPATH").foreach { value =>
|
sys.env.get("SPARK_CLASSPATH").foreach { value =>
|
||||||
val error =
|
val warning =
|
||||||
s"""
|
s"""
|
||||||
|SPARK_CLASSPATH was detected (set to '$value').
|
|SPARK_CLASSPATH was detected (set to '$value').
|
||||||
| This has undefined behavior when running on a cluster and is deprecated in Spark 1.0+.
|
|This is deprecated in Spark 1.0+.
|
||||||
|
|
|
|
||||||
|Please instead use:
|
|Please instead use:
|
||||||
| - ./spark-submit with --driver-class-path to augment the driver classpath
|
| - ./spark-submit with --driver-class-path to augment the driver classpath
|
||||||
| - spark.executor.extraClassPath to augment the executor classpath
|
| - spark.executor.extraClassPath to augment the executor classpath
|
||||||
""".stripMargin
|
""".stripMargin
|
||||||
logError(error)
|
logWarning(warning)
|
||||||
|
|
||||||
for (key <- Seq(executorClasspathKey, driverClassPathKey)) {
|
for (key <- Seq(executorClasspathKey, driverClassPathKey)) {
|
||||||
if (getOption(key).isDefined) {
|
if (getOption(key).isDefined) {
|
||||||
|
|
Loading…
Reference in a new issue