spark-instrumented-optimizer/project
Shixiong Zhu 4d7c359263 [SPARK-14437][CORE] Use the address that NettyBlockTransferService listens to create BlockManagerId
## What changes were proposed in this pull request?

Here is why SPARK-14437 happens:
BlockManagerId is created using NettyBlockTransferService.hostName which comes from `customHostname`. And `Executor` will set `customHostname` to the hostname which is detected by the driver. However, the driver may not be able to detect the correct address in some complicated network (Netty's Channel.remoteAddress doesn't always return a connectable address). In such case, `BlockManagerId` will be created using a wrong hostname.

To fix this issue, this PR uses `hostname` provided by `SparkEnv.create` to create `NettyBlockTransferService` and set `NettyBlockTransferService.hostname` to this one directly. A bonus of this approach is NettyBlockTransferService won't bound to `0.0.0.0` which is much safer.

## How was this patch tested?

Manually checked the bound address using local-cluster.

Author: Shixiong Zhu <shixiong@databricks.com>

Closes #12240 from zsxwing/SPARK-14437.
2016-04-08 17:18:19 -07:00
..
project [MINOR][BUILD] Changed the comment to reflect the plugin project is there to support SBT pom reader only. 2015-11-30 09:30:58 +00:00
build.properties [SPARK-13834][BUILD] Update sbt and sbt plugins for 2.x. 2016-03-13 18:47:04 -07:00
MimaBuild.scala [SPARK-13874][DOC] Remove docs of streaming-akka, streaming-zeromq, streaming-mqtt and streaming-twitter 2016-03-26 01:47:27 -07:00
MimaExcludes.scala [SPARK-14437][CORE] Use the address that NettyBlockTransferService listens to create BlockManagerId 2016-04-08 17:18:19 -07:00
plugins.sbt [SPARK-14366] Remove sbt-idea plugin 2016-04-04 16:55:59 -07:00
SparkBuild.scala [SPARK-529][SQL] Modify SQLConf to use new config API from core. 2016-04-05 15:19:51 -07:00