spark-instrumented-optimizer/docs
Marcelo Vanzin 20a6013106 [SPARK-2996] Implement userClassPathFirst for driver, yarn.
Yarn's config option `spark.yarn.user.classpath.first` does not work the same way as
`spark.files.userClassPathFirst`; Yarn's version is a lot more dangerous, in that it
modifies the system classpath, instead of restricting the changes to the user's class
loader. So this change implements the behavior of the latter for Yarn, and deprecates
the more dangerous choice.

To be able to achieve feature-parity, I also implemented the option for drivers (the existing
option only applies to executors). So now there are two options, each controlling whether
to apply userClassPathFirst to the driver or executors. The old option was deprecated, and
aliased to the new one (`spark.executor.userClassPathFirst`).

The existing "child-first" class loader also had to be fixed. It didn't handle resources, and it
was also doing some things that ended up causing JVM errors depending on how things
were being called.

Author: Marcelo Vanzin <vanzin@cloudera.com>

Closes #3233 from vanzin/SPARK-2996 and squashes the following commits:

9cf9cf1 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
a1499e2 [Marcelo Vanzin] Remove SPARK_HOME propagation.
fa7df88 [Marcelo Vanzin] Remove 'test.resource' file, create it dynamically.
a8c69f1 [Marcelo Vanzin] Review feedback.
cabf962 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
a1b8d7e [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
3f768e3 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
2ce3c7a [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
0e6d6be [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
70d4044 [Marcelo Vanzin] Fix pyspark/yarn-cluster test.
0fe7777 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
0e6ef19 [Marcelo Vanzin] Move class loaders around and make names more meaninful.
fe970a7 [Marcelo Vanzin] Review feedback.
25d4fed [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
3cb6498 [Marcelo Vanzin] Call the right loadClass() method on the parent.
fbb8ab5 [Marcelo Vanzin] Add locking in loadClass() to avoid deadlocks.
2e6c4b7 [Marcelo Vanzin] Mention new setting in documentation.
b6497f9 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
a10f379 [Marcelo Vanzin] Some feedback.
3730151 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
f513871 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
44010b6 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
7b57cba [Marcelo Vanzin] Remove now outdated message.
5304d64 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
35949c8 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
54e1a98 [Marcelo Vanzin] Merge branch 'master' into SPARK-2996
d1273b2 [Marcelo Vanzin] Add test file to rat exclude.
fa1aafa [Marcelo Vanzin] Remove write check on user jars.
89d8072 [Marcelo Vanzin] Cleanups.
a963ea3 [Marcelo Vanzin] Implement spark.driver.userClassPathFirst for standalone cluster mode.
50afa5f [Marcelo Vanzin] Fix Yarn executor command line.
7d14397 [Marcelo Vanzin] Register user jars in executor up front.
7f8603c [Marcelo Vanzin] Fix yarn-cluster mode without userClassPathFirst.
20373f5 [Marcelo Vanzin] Fix ClientBaseSuite.
55c88fa [Marcelo Vanzin] Run all Yarn integration tests via spark-submit.
0b64d92 [Marcelo Vanzin] Add deprecation warning to yarn option.
4a84d87 [Marcelo Vanzin] Fix the child-first class loader.
d0394b8 [Marcelo Vanzin] Add "deprecated configs" to SparkConf.
46d8cf2 [Marcelo Vanzin] Update doc with new option, change name to "userClassPathFirst".
a314f2d [Marcelo Vanzin] Enable driver class path isolation in SparkSubmit.
91f7e54 [Marcelo Vanzin] [yarn] Enable executor class path isolation.
a853e74 [Marcelo Vanzin] Re-work CoarseGrainedExecutorBackend command line arguments.
89522ef [Marcelo Vanzin] Add class path isolation support for Yarn cluster mode.
2015-02-09 21:17:28 -08:00
..
_layouts [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
_plugins [SPARK-4501][Core] - Create build/mvn to automatically download maven/zinc/scalac 2014-12-27 13:26:38 -08:00
css [SPARK-1566] consolidate programming guide, and general doc updates 2014-05-30 00:34:33 -07:00
img [SPARK-4259][MLlib]: Add Power Iteration Clustering Algorithm with Gaussian Similarity Function 2015-01-30 14:09:49 -08:00
js [SPARK-1566] consolidate programming guide, and general doc updates 2014-05-30 00:34:33 -07:00
_config.yml [DOC] Fixed Mesos version in doc from 0.18.1 to 0.21.0 2015-01-09 14:40:45 -08:00
api.md [SPARK-1439, SPARK-1440] Generate unified Scaladoc across projects and Javadocs 2014-04-21 21:57:40 -07:00
bagel-programming-guide.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
building-spark.md [Doc] Minor: Fixes several formatting issues 2015-02-02 21:14:21 -08:00
cluster-overview.md [SPARK-1566] consolidate programming guide, and general doc updates 2014-05-30 00:34:33 -07:00
configuration.md [SPARK-2996] Implement userClassPathFirst for driver, yarn. 2015-02-09 21:17:28 -08:00
contributing-to-spark.md Work in progress: 2013-09-08 00:29:11 -07:00
ec2-scripts.md Update ec2-scripts.md 2015-02-06 11:04:48 +00:00
graphx-programming-guide.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
hadoop-third-party-distributions.md [SPARK-4501][Core] - Create build/mvn to automatically download maven/zinc/scalac 2014-12-27 13:26:38 -08:00
hardware-provisioning.md Change port from 3030 to 4040 2013-09-11 10:01:38 -07:00
index.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
java-programming-guide.md [SPARK-1566] consolidate programming guide, and general doc updates 2014-05-30 00:34:33 -07:00
job-scheduling.md SPARK-4585. Spark dynamic executor allocation should use minExecutors as... 2015-02-02 12:27:08 -08:00
ml-guide.md [SPARK-5254][MLLIB] remove developers section from spark.ml guide 2015-01-14 18:54:17 -08:00
mllib-classification-regression.md SPARK-2830 [MLlib]: re-organize mllib documentation 2014-08-12 17:15:21 -07:00
mllib-clustering.md [SPARK-5539][MLLIB] LDA guide 2015-02-08 23:40:36 -08:00
mllib-collaborative-filtering.md [MLlib] fix python example of ALS in guide 2015-01-27 15:33:01 -08:00
mllib-data-types.md [SPARK-2830][MLLIB] doc update for 1.1 2014-08-27 01:19:48 -07:00
mllib-decision-tree.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
mllib-dimensionality-reduction.md SPARK-1307 [DOCS] Don't use term 'standalone' to refer to a Spark Application 2014-10-14 21:37:51 -07:00
mllib-ensembles.md [SPARK-4580] [SPARK-4610] [mllib] [docs] Documentation for tree ensembles + DecisionTree API fix 2014-12-04 09:57:50 +08:00
mllib-feature-extraction.md [SPARK-5207] [MLLIB] StandardScalerModel mean and variance re-use 2015-02-01 09:21:14 -08:00
mllib-guide.md SPARK-5665 [DOCS] Update netlib-java documentation 2015-02-08 16:34:26 -08:00
mllib-linear-methods.md [SPARK-4711] [mllib] [docs] Programming guide advice on choosing optimizer 2014-12-04 08:58:03 +08:00
mllib-naive-bayes.md fixed MLlib Naive-Bayes java example bug 2014-11-04 09:53:43 -08:00
mllib-optimization.md [SPARK-4711] [mllib] [docs] Programming guide advice on choosing optimizer 2014-12-04 08:58:03 +08:00
mllib-statistics.md [SPARK-3964] [MLlib] [PySpark] add Hypothesis test Python API 2014-11-04 21:35:52 -08:00
monitoring.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
programming-guide.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
python-programming-guide.md [SPARK-1566] consolidate programming guide, and general doc updates 2014-05-30 00:34:33 -07:00
quick-start.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
README.md [SPARK-4501][Core] - Create build/mvn to automatically download maven/zinc/scalac 2014-12-27 13:26:38 -08:00
running-on-mesos.md Add mesos specific configurations into doc 2014-12-18 12:15:53 -08:00
running-on-yarn.md [SPARK-2945][YARN][Doc]add doc for spark.executor.instances 2015-02-06 11:58:22 -08:00
scala-programming-guide.md [SPARK-1566] consolidate programming guide, and general doc updates 2014-05-30 00:34:33 -07:00
security.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
spark-standalone.md [SPARK-4771][Docs] Document standalone cluster supervise mode 2014-12-10 12:41:36 -08:00
sql-programming-guide.md [Branch-1.3] [DOC] doc fix for date 2015-02-05 12:42:27 -08:00
storage-openstack-swift.md [SPARK-938][doc] Add OpenStack Swift support 2014-09-07 20:56:04 -07:00
streaming-custom-receivers.md [SPARK-4806] Streaming doc update for 1.2 2014-12-11 06:21:23 -08:00
streaming-flume-integration.md [SPARK-4806] Streaming doc update for 1.2 2014-12-11 06:21:23 -08:00
streaming-kafka-integration.md [SPARK-5058] Part 2. Typos and broken URL 2015-01-23 23:34:11 -08:00
streaming-kinesis-integration.md SPARK-3069 [DOCS] Build instructions in README are outdated 2014-09-16 09:18:03 -07:00
streaming-programming-guide.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00
submitting-applications.md SPARK-4506 [DOCS] Addendum: Update more docs to reflect that standalone works in cluster mode 2015-01-25 15:25:05 -08:00
tuning.md [SPARK-5608] Improve SEO of Spark documentation pages 2015-02-05 11:12:50 -08:00

Welcome to the Spark documentation!

This readme will walk you through navigating and building the Spark documentation, which is included here with the Spark source code. You can also find documentation specific to release versions of Spark at http://spark.apache.org/documentation.html.

Read on to learn more about viewing documentation in plain text (i.e., markdown) or building the documentation yourself. Why build it yourself? So that you have the docs that corresponds to whichever version of Spark you currently have checked out of revision control.

Generating the Documentation HTML

We include the Spark documentation as part of the source (as opposed to using a hosted wiki, such as the github wiki, as the definitive documentation) to enable the documentation to evolve along with the source code and be captured by revision control (currently git). This way the code automatically includes the version of the documentation that is relevant regardless of which version or release you have checked out or downloaded.

In this directory you will find textfiles formatted using Markdown, with an ".md" suffix. You can read those text files directly if you want. Start with index.md.

The markdown code can be compiled to HTML using the Jekyll tool. Jekyll and a few dependencies must be installed for this to work. We recommend installing via the Ruby Gem dependency manager. Since the exact HTML output varies between versions of Jekyll and its dependencies, we list specific versions here in some cases:

$ sudo gem install jekyll
$ sudo gem install jekyll-redirect-from

Execute jekyll from the docs/ directory. Compiling the site with Jekyll will create a directory called _site containing index.html as well as the rest of the compiled files.

You can modify the default Jekyll build as follows:

# Skip generating API docs (which takes a while)
$ SKIP_API=1 jekyll build
# Serve content locally on port 4000
$ jekyll serve --watch
# Build the site with extra features used on the live page
$ PRODUCTION=1 jekyll build

Pygments

We also use pygments (http://pygments.org) for syntax highlighting in documentation markdown pages, so you will also need to install that (it requires Python) by running sudo pip install Pygments.

To mark a block of code in your markdown to be syntax highlighted by jekyll during the compile phase, use the following sytax:

{% highlight scala %}
// Your scala code goes here, you can replace scala with many other
// supported languages too.
{% endhighlight %}

Sphinx

We use Sphinx to generate Python API docs, so you will need to install it by running sudo pip install sphinx.

API Docs (Scaladoc and Sphinx)

You can build just the Spark scaladoc by running build/sbt doc from the SPARK_PROJECT_ROOT directory.

Similarly, you can build just the PySpark docs by running make html from the SPARK_PROJECT_ROOT/python/docs directory. Documentation is only generated for classes that are listed as public in __init__.py.

When you run jekyll in the docs directory, it will also copy over the scaladoc for the various Spark subprojects into the docs directory (and then also into the _site directory). We use a jekyll plugin to run build/sbt doc before building the site so if you haven't run it (recently) it may take some time as it generates all of the scaladoc. The jekyll plugin also generates the PySpark docs Sphinx.

NOTE: To skip the step of building and copying over the Scala and Python API docs, run SKIP_API=1 jekyll.