spark-instrumented-optimizer/python
Aaron Davidson 0307db0f55 SPARK-1099: Introduce local[*] mode to infer number of cores
This is the default mode for running spark-shell and pyspark, intended to allow users running spark for the first time to see the performance benefits of using multiple cores, while not breaking backwards compatibility for users who use "local" mode and expect exactly 1 core.

Author: Aaron Davidson <aaron@databricks.com>

Closes #182 from aarondav/110 and squashes the following commits:

a88294c [Aaron Davidson] Rebased changes for new spark-shell
a9f393e [Aaron Davidson] SPARK-1099: Introduce local[*] mode to infer number of cores
2014-04-07 13:06:30 -07:00
..
examples Merge pull request #562 from jyotiska/master. Closes #562. 2014-02-08 23:36:48 -08:00
lib Changes on top of Prashant's patch. 2014-01-03 18:30:17 -08:00
pyspark SPARK-1099: Introduce local[*] mode to infer number of cores 2014-04-07 13:06:30 -07:00
test_support License headers 2013-12-09 16:41:01 -08:00
.gitignore Rename top-level 'pyspark' directory to 'python' 2013-01-01 15:05:00 -08:00
epydoc.conf Fix some Python docs and make sure to unset SPARK_TESTING in Python 2013-12-29 20:15:07 -05:00
run-tests SPARK-1336 Reducing the output of run-tests script. 2014-03-29 23:03:03 -07:00