spark-instrumented-optimizer/python/docs/source/development/testing.rst
Josh Soref 13fd272cd3 Spelling r common dev mlib external project streaming resource managers python
### What changes were proposed in this pull request?

This PR intends to fix typos in the sub-modules:
* `R`
* `common`
* `dev`
* `mlib`
* `external`
* `project`
* `streaming`
* `resource-managers`
* `python`

Split per srowen https://github.com/apache/spark/pull/30323#issuecomment-728981618

NOTE: The misspellings have been reported at 706a726f87 (commitcomment-44064356)

### Why are the changes needed?

Misspelled words make it harder to read / understand content.

### Does this PR introduce _any_ user-facing change?

There are various fixes to documentation, etc...

### How was this patch tested?

No testing was performed

Closes #30402 from jsoref/spelling-R_common_dev_mlib_external_project_streaming_resource-managers_python.

Authored-by: Josh Soref <jsoref@users.noreply.github.com>
Signed-off-by: Sean Owen <srowen@gmail.com>
2020-11-27 10:22:45 -06:00

58 lines
2.1 KiB
ReStructuredText

.. Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements. See the NOTICE file
distributed with this work for additional information
regarding copyright ownership. The ASF licenses this file
to you under the Apache License, Version 2.0 (the
"License"); you may not use this file except in compliance
with the License. You may obtain a copy of the License at
.. http://www.apache.org/licenses/LICENSE-2.0
.. Unless required by applicable law or agreed to in writing,
software distributed under the License is distributed on an
"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied. See the License for the
specific language governing permissions and limitations
under the License.
===============
Testing PySpark
===============
In order to run PySpark tests, you should build Spark itself first via Maven or SBT. For example,
.. code-block:: bash
build/mvn -DskipTests clean package
After that, the PySpark test cases can be run via using ``python/run-tests``. For example,
.. code-block:: bash
python/run-tests --python-executable=python3
Note that you may set ``OBJC_DISABLE_INITIALIZE_FORK_SAFETY`` environment variable to ``YES`` if you are running tests on Mac OS.
Please see the guidance on how to `build Spark <https://github.com/apache/spark#building-spark>`_,
`run tests for a module, or individual tests <https://spark.apache.org/developer-tools.html>`_.
Running Individual PySpark Tests
--------------------------------
You can run a specific test via using ``python/run-tests``, for example, as below:
.. code-block:: bash
python/run-tests --testnames pyspark.sql.tests.test_arrow
Please refer to `Testing PySpark <https://spark.apache.org/developer-tools.html>`_ for more details.
Running tests using GitHub Actions
----------------------------------
You can run the full PySpark tests by using GitHub Actions in your own forked GitHub
repository with a few clicks. Please refer to
`Running tests in your forked repository using GitHub Actions <https://spark.apache.org/developer-tools.html>`_ for more details.