[SPARK-35562][DOC] Fix docs about Kubernetes and Yarn
Fixed some places in cluster-overview that are obsolete (i.e. not mentioning Kubernetes), and also fixed the Yarn spark-submit sample command in submitting-applications. ### What changes were proposed in this pull request? This is to fix the docs in "Cluster Overview" and "Submitting Applications" for places where Kubernetes is missed (mostly due to obsolete docs that haven't got updated) and where Yarn sample spark-submit command is incorrectly written. ### Why are the changes needed? To help the Spark users who uses Kubernetes as cluster manager to have a correct idea when reading the "Cluster Overview" doc page. Also to make the sample spark-submit command for Yarn actually runnable in the "Submitting Applications" doc page, by removing the invalid comment after line continuation char `\`. ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? No test, as this is doc fix. Closes #32701 from huskysun/doc-fix. Authored-by: Shiqi Sun <s.sun@salesforce.com> Signed-off-by: Dongjoon Hyun <dhyun@apple.com>
This commit is contained in:
parent
73ba4492b1
commit
8c69e9cd94
|
@ -28,7 +28,7 @@ Spark applications run as independent sets of processes on a cluster, coordinate
|
|||
object in your main program (called the _driver program_).
|
||||
|
||||
Specifically, to run on a cluster, the SparkContext can connect to several types of _cluster managers_
|
||||
(either Spark's own standalone cluster manager, Mesos or YARN), which allocate resources across
|
||||
(either Spark's own standalone cluster manager, Mesos, YARN or Kubernetes), which allocate resources across
|
||||
applications. Once connected, Spark acquires *executors* on nodes in the cluster, which are
|
||||
processes that run computations and store data for your application.
|
||||
Next, it sends your application code (defined by JAR or Python files passed to SparkContext) to
|
||||
|
@ -48,7 +48,7 @@ There are several useful things to note about this architecture:
|
|||
writing it to an external storage system.
|
||||
2. Spark is agnostic to the underlying cluster manager. As long as it can acquire executor
|
||||
processes, and these communicate with each other, it is relatively easy to run it even on a
|
||||
cluster manager that also supports other applications (e.g. Mesos/YARN).
|
||||
cluster manager that also supports other applications (e.g. Mesos/YARN/Kubernetes).
|
||||
3. The driver program must listen for and accept incoming connections from its executors throughout
|
||||
its lifetime (e.g., see [spark.driver.port in the network config
|
||||
section](configuration.html#networking)). As such, the driver program must be network
|
||||
|
@ -117,7 +117,7 @@ The following table summarizes terms you'll see used to refer to cluster concept
|
|||
</tr>
|
||||
<tr>
|
||||
<td>Cluster manager</td>
|
||||
<td>An external service for acquiring resources on the cluster (e.g. standalone manager, Mesos, YARN)</td>
|
||||
<td>An external service for acquiring resources on the cluster (e.g. standalone manager, Mesos, YARN, Kubernetes)</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>Deploy mode</td>
|
||||
|
|
|
@ -114,12 +114,12 @@ run it with `--help`. Here are a few examples of common options:
|
|||
/path/to/examples.jar \
|
||||
1000
|
||||
|
||||
# Run on a YARN cluster
|
||||
# Run on a YARN cluster in cluster deploy mode
|
||||
export HADOOP_CONF_DIR=XXX
|
||||
./bin/spark-submit \
|
||||
--class org.apache.spark.examples.SparkPi \
|
||||
--master yarn \
|
||||
--deploy-mode cluster \ # can be client for client mode
|
||||
--deploy-mode cluster \
|
||||
--executor-memory 20G \
|
||||
--num-executors 50 \
|
||||
/path/to/examples.jar \
|
||||
|
|
Loading…
Reference in a new issue