Apache Spark - A unified analytics engine for large-scale data processing
Go to file
Sean Owen 8a17d26784 [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials
## What changes were proposed in this pull request?

I want to get rid of as much use of `scala.language.existentials` as possible for 3.0. It's a complicated language feature that generates warnings unless this value is imported. It might even be on the way out of Scala: https://contributors.scala-lang.org/t/proposal-to-remove-existential-types-from-the-language/2785

For Spark, it comes up mostly where the code plays fast and loose with generic types, not the advanced situations you'll often see referenced where this feature is explained. For example, it comes up in cases where a function returns something like `(String, Class[_])`. Scala doesn't like matching this to any other instance of `(String, Class[_])` because doing so requires inferring the existence of some type that satisfies both. Seems obvious if the generic type is a wildcard, but, not technically something Scala likes to let you get away with.

This is a large PR, and it only gets rid of _most_ instances of `scala.language.existentials`. The change should be all compile-time and shouldn't affect APIs or logic.

Many of the changes simply touch up sloppiness about generic types, making the known correct value explicit in the code.

Some fixes involve being more explicit about the existence of generic types in methods. For instance, `def foo(arg: Class[_])` seems innocent enough but should really be declared `def foo[T](arg: Class[T])` to let Scala select and fix a single type when evaluating calls to `foo`.

For kind of surprising reasons, this comes up in places where code evaluates a tuple of things that involve a generic type, but is OK if the two parts of the tuple are evaluated separately.

One key change was altering `Utils.classForName(...): Class[_]` to the more correct `Utils.classForName[T](...): Class[T]`. This caused a number of small but positive changes to callers that otherwise had to cast the result.

In several tests, `Dataset[_]` was used where `DataFrame` seems to be the clear intent.

Finally, in a few cases in MLlib, the return type `this.type` was used where there are no subclasses of the class that uses it. This really isn't needed and causes issues for Scala reasoning about the return type. These are just changed to be concrete classes as return types.

After this change, we have only a few classes that still import `scala.language.existentials` (because modifying them would require extensive rewrites to fix) and no build warnings.

## How was this patch tested?

Existing tests.

Closes #24431 from srowen/SPARK-27536.

Authored-by: Sean Owen <sean.owen@databricks.com>
Signed-off-by: Sean Owen <sean.owen@databricks.com>
2019-04-29 11:02:01 -05:00
.github [SPARK-18073][DOCS][WIP] Migrate wiki to spark.apache.org web site 2016-11-23 11:25:47 +00:00
assembly [SPARK-26134][CORE] Upgrading Hadoop to 2.7.4 to fix java.version problem 2018-11-21 23:09:57 -08:00
bin [SPARK-26132][BUILD][CORE] Remove support for Scala 2.11 in Spark 3.0.0 2019-03-25 10:46:42 -05:00
build [SPARK-26144][BUILD] build/mvn should detect scala.version based on scala.binary.version 2018-11-22 14:49:41 -08:00
common [SPARK-27404][CORE][SQL][STREAMING][YARN] Fix build warnings for 3.0: postfixOps edition 2019-04-11 13:43:44 -05:00
conf [SPARK-26890][DOC] Add list of available Dropwizard metrics in Spark and add additional configuration details to the monitoring documentation 2019-02-27 10:07:15 -06:00
core [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials 2019-04-29 11:02:01 -05:00
data [SPARK-22666][ML][SQL] Spark datasource for image format 2018-09-05 11:59:00 -07:00
dev Revert "[SPARK-27467][FOLLOW-UP][BUILD] Upgrade Maven to 3.6.1 in AppVeyor and Doc" 2019-04-28 11:03:15 +09:00
docs [SPARK-27472] add user guide for binary file data source 2019-04-29 08:58:56 -07:00
examples [SPARK-26970][PYTHON][ML] Add Spark ML interaction transformer to PySpark 2019-04-23 13:53:33 -07:00
external [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials 2019-04-29 11:02:01 -05:00
graphx [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials 2019-04-29 11:02:01 -05:00
hadoop-cloud [SPARK-27175][BUILD] Upgrade hadoop-3 to 3.2.0 2019-03-16 19:42:05 -05:00
launcher [SPARK-27397][CORE] Take care of OpenJ9 JVM in Spark 2019-04-16 09:11:47 -05:00
licenses [SPARK-27358][UI] Update jquery to 1.12.x to pick up security fixes 2019-04-05 12:54:01 -05:00
licenses-binary [SPARK-27358][UI] Update jquery to 1.12.x to pick up security fixes 2019-04-05 12:54:01 -05:00
mllib [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials 2019-04-29 11:02:01 -05:00
mllib-local [SPARK-19591][ML][MLLIB] Add sample weights to decision trees 2019-01-24 18:20:28 -07:00
project [MINOR][BUILD] Update genjavadoc to 0.13 2019-04-24 13:44:48 +09:00
python [SPARK-23619][DOCS] Add output description for some generator expressions / functions 2019-04-27 10:30:12 +09:00
R [SPARK-23619][DOCS] Add output description for some generator expressions / functions 2019-04-27 10:30:12 +09:00
repl [SPARK-27323][CORE][SQL][STREAMING] Use Single-Abstract-Method support in Scala 2.12 to simplify code 2019-04-02 07:37:05 -07:00
resource-managers [SPARK-26729][K8S] Fix typo with default value for R image name 2019-04-24 21:08:42 -07:00
sbin [SPARK-27056][MESOS] Remove start-shuffle-service.sh 2019-03-08 18:51:38 -06:00
sql [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials 2019-04-29 11:02:01 -05:00
streaming [SPARK-27536][CORE][ML][SQL][STREAMING] Remove most use of scala.language.existentials 2019-04-29 11:02:01 -05:00
tools [SPARK-25956] Make Scala 2.12 as default Scala version in Spark 3.0 2018-11-14 16:22:23 -08:00
.gitattributes [SPARK-3870] EOL character enforcement 2014-10-31 12:39:52 -07:00
.gitignore [MINOR][DOC] Documentation on JVM options for SBT 2019-01-22 18:27:24 -06:00
appveyor.yml [MINOR][BUILD] Remove -Phive-thriftserver profile within appveyor.yml 2018-07-30 10:01:18 +08:00
CONTRIBUTING.md [SPARK-18073][DOCS][WIP] Migrate wiki to spark.apache.org web site 2016-11-23 11:25:47 +00:00
LICENSE [SPARK-24654][BUILD] Update, fix LICENSE and NOTICE, and specialize for source vs binary 2018-06-30 19:27:16 -05:00
LICENSE-binary [SPARK-27469][CORE] Update Commons BeanUtils to 1.9.3 2019-04-15 19:18:37 -07:00
NOTICE [SPARK-23654][BUILD] remove jets3t as a dependency of spark 2018-08-16 12:34:23 -07:00
NOTICE-binary [SPARK-27054][BUILD][SQL] Remove the Calcite dependency 2019-03-09 16:34:24 -08:00
pom.xml Revert "[SPARK-27467][BUILD][TEST-MAVEN] Upgrade Maven to 3.6.1" 2019-04-28 11:03:04 +09:00
README.md [SPARK-7721][INFRA] Run and generate test coverage report from Python via Jenkins 2019-02-01 10:18:08 +08:00
scalastyle-config.xml [SPARK-25986][BUILD] Add rules to ban throw Errors in application code 2018-11-14 13:05:18 -08:00

Apache Spark

Jenkins Build AppVeyor Build PySpark Coverage

Spark is a fast and general cluster computing system for Big Data. It provides high-level APIs in Scala, Java, Python, and R, and an optimized engine that supports general computation graphs for data analysis. It also supports a rich set of higher-level tools including Spark SQL for SQL and DataFrames, MLlib for machine learning, GraphX for graph processing, and Spark Streaming for stream processing.

http://spark.apache.org/

Online Documentation

You can find the latest Spark documentation, including a programming guide, on the project web page. This README file only contains basic setup instructions.

Building Spark

Spark is built using Apache Maven. To build Spark and its example programs, run:

build/mvn -DskipTests clean package

(You do not need to do this if you downloaded a pre-built package.)

You can build Spark using more than one thread by using the -T option with Maven, see "Parallel builds in Maven 3". More detailed documentation is available from the project site, at "Building Spark".

For general development tips, including info on developing Spark using an IDE, see "Useful Developer Tools".

Interactive Scala Shell

The easiest way to start using Spark is through the Scala shell:

./bin/spark-shell

Try the following command, which should return 1000:

scala> sc.parallelize(1 to 1000).count()

Interactive Python Shell

Alternatively, if you prefer Python, you can use the Python shell:

./bin/pyspark

And run the following command, which should also return 1000:

>>> sc.parallelize(range(1000)).count()

Example Programs

Spark also comes with several sample programs in the examples directory. To run one of them, use ./bin/run-example <class> [params]. For example:

./bin/run-example SparkPi

will run the Pi example locally.

You can set the MASTER environment variable when running examples to submit examples to a cluster. This can be a mesos:// or spark:// URL, "yarn" to run on YARN, and "local" to run locally with one thread, or "local[N]" to run locally with N threads. You can also use an abbreviated class name if the class is in the examples package. For instance:

MASTER=spark://host:7077 ./bin/run-example SparkPi

Many of the example programs print usage help if no params are given.

Running Tests

Testing first requires building Spark. Once Spark is built, tests can be run using:

./dev/run-tests

Please see the guidance on how to run tests for a module, or individual tests.

There is also a Kubernetes integration test, see resource-managers/kubernetes/integration-tests/README.md

A Note About Hadoop Versions

Spark uses the Hadoop core library to talk to HDFS and other Hadoop-supported storage systems. Because the protocols have changed in different versions of Hadoop, you must build Spark against the same version that your cluster runs.

Please refer to the build documentation at "Specifying the Hadoop Version and Enabling YARN" for detailed guidance on building for a particular distribution of Hadoop, including building for particular Hive and Hive Thriftserver distributions.

Configuration

Please refer to the Configuration Guide in the online documentation for an overview on how to configure Spark.

Contributing

Please review the Contribution to Spark guide for information on how to get started contributing to the project.