spark-instrumented-optimizer/python
Xiangrui Meng 94c6c06ea1 [FIX] do not load defaults when testing SparkConf in pyspark
The default constructor loads default properties, which can fail the test.

Author: Xiangrui Meng <meng@databricks.com>

Closes #775 from mengxr/pyspark-conf-fix and squashes the following commits:

83ef6c4 [Xiangrui Meng] do not load defaults when testing SparkConf in pyspark
2014-05-14 14:57:17 -07:00
..
lib SPARK-1004. PySpark on YARN 2014-04-29 23:24:34 -07:00
pyspark [FIX] do not load defaults when testing SparkConf in pyspark 2014-05-14 14:57:17 -07:00
test_support License headers 2013-12-09 16:41:01 -08:00
.gitignore SPARK-1004. PySpark on YARN 2014-04-29 23:24:34 -07:00
epydoc.conf [SPARK-1439, SPARK-1440] Generate unified Scaladoc across projects and Javadocs 2014-04-21 21:57:40 -07:00
run-tests FIX: Don't build Hive in assembly unless running Hive tests. 2014-04-17 17:24:00 -07:00