refactor sparkHome to val
clean code
This commit is contained in:
parent
29f4b6a2d9
commit
2b3c461451
|
@ -76,7 +76,7 @@ private[spark] class Worker(
|
|||
@volatile var registered = false
|
||||
@volatile var connected = false
|
||||
val workerId = generateWorkerId()
|
||||
var sparkHome: File = null
|
||||
val sparkHome = new File(Option(System.getenv("SPARK_HOME")).getOrElse("."))
|
||||
var workDir: File = null
|
||||
val executors = new HashMap[String, ExecutorRunner]
|
||||
val finishedExecutors = new HashMap[String, ExecutorRunner]
|
||||
|
@ -120,7 +120,6 @@ private[spark] class Worker(
|
|||
assert(!registered)
|
||||
logInfo("Starting Spark worker %s:%d with %d cores, %s RAM".format(
|
||||
host, port, cores, Utils.megabytesToString(memory)))
|
||||
sparkHome = new File(Option(System.getenv("SPARK_HOME")).getOrElse("."))
|
||||
logInfo("Spark home: " + sparkHome)
|
||||
createWorkDir()
|
||||
webUi = new WorkerWebUI(this, workDir, Some(webUiPort))
|
||||
|
@ -211,7 +210,8 @@ private[spark] class Worker(
|
|||
try {
|
||||
logInfo("Asked to launch executor %s/%d for %s".format(appId, execId, appDesc.name))
|
||||
val manager = new ExecutorRunner(appId, execId, appDesc, cores_, memory_,
|
||||
self, workerId, host, new File(appDesc.sparkHome.getOrElse(sparkHome.getAbsolutePath)),
|
||||
self, workerId, host,
|
||||
appDesc.sparkHome.map(userSparkHome => new File(userSparkHome)).getOrElse(sparkHome),
|
||||
workDir, akkaUrl, ExecutorState.RUNNING)
|
||||
executors(appId + "/" + execId) = manager
|
||||
manager.start()
|
||||
|
@ -225,6 +225,7 @@ private[spark] class Worker(
|
|||
logError("Failed to launch exector %s/%d for %s".format(appId, execId, appDesc.name))
|
||||
if (executors.contains(appId + "/" + execId)) {
|
||||
executors(appId + "/" + execId).kill()
|
||||
executors -= appId + "/" + execId
|
||||
}
|
||||
masterLock.synchronized {
|
||||
master ! ExecutorStateChanged(appId, execId, ExecutorState.FAILED, None, None)
|
||||
|
|
Loading…
Reference in a new issue