spark-instrumented-optimizer/repl
Wenchen Fan a432a2b860 [SPARK-15116] In REPL we should create SparkSession first and get SparkContext from it
## What changes were proposed in this pull request?

see https://github.com/apache/spark/pull/12873#discussion_r61993910. The problem is, if we create `SparkContext` first and then call `SparkSession.builder.enableHiveSupport().getOrCreate()`, we will reuse the existing `SparkContext` and the hive flag won't be set.

## How was this patch tested?

verified it locally.

Author: Wenchen Fan <wenchen@databricks.com>

Closes #12890 from cloud-fan/repl.
2016-05-04 14:40:54 -07:00
..
scala-2.10/src [SPARK-15116] In REPL we should create SparkSession first and get SparkContext from it 2016-05-04 14:40:54 -07:00
scala-2.11/src [SPARK-15116] In REPL we should create SparkSession first and get SparkContext from it 2016-05-04 14:40:54 -07:00
src [SPARK-10001] Consolidate Signaling and SignalLogger. 2016-04-22 09:36:59 -07:00
pom.xml Revert "[SPARK-14613][ML] Add @Since into the matrix and vector classes in spark-mllib-local" 2016-04-28 19:57:41 -07:00