spark-instrumented-optimizer/python/pyspark
Shivaram Venkataraman d3234f9726 Make collectPartitions take an array of partitions
Change the implementation to use runJob instead of PartitionPruningRDD.
Also update the unit tests and the python take implementation
to use the new interface.
2013-12-19 11:40:34 -08:00
..
__init__.py Export StorageLevel and refactor 2013-09-07 14:41:31 -07:00
accumulators.py Add custom serializer support to PySpark. 2013-11-10 16:45:38 -08:00
broadcast.py Add Apache license headers and LICENSE and NOTICE files 2013-07-16 17:21:33 -07:00
cloudpickle.py Rename top-level 'pyspark' directory to 'python' 2013-01-01 15:05:00 -08:00
context.py Add collectPartition to JavaRDD interface. 2013-12-18 11:40:07 -08:00
daemon.py Add Apache license headers and LICENSE and NOTICE files 2013-07-16 17:21:33 -07:00
files.py Initial work to rename package to org.apache.spark 2013-09-01 14:13:13 -07:00
java_gateway.py Further fixes to get PySpark to work on Windows 2013-09-02 01:19:29 +00:00
join.py Change numSplits to numPartitions in PySpark. 2013-02-24 13:25:09 -08:00
rdd.py Make collectPartitions take an array of partitions 2013-12-19 11:40:34 -08:00
rddsampler.py RDD sample() and takeSample() prototypes for PySpark 2013-08-28 16:46:13 -07:00
serializers.py Removed unused basestring case from dump_stream. 2013-11-26 14:35:12 -08:00
shell.py Update build version in master 2013-09-24 11:41:51 -07:00
statcounter.py Implementing SPARK-838: Add DoubleRDDFunctions methods to PySpark 2013-08-21 17:05:58 -07:00
storagelevel.py Export StorageLevel and refactor 2013-09-07 14:41:31 -07:00
tests.py Fix UnicodeEncodeError in PySpark saveAsTextFile(). 2013-11-28 23:44:56 -08:00
worker.py FramedSerializer: _dumps => dumps, _loads => loads. 2013-11-10 17:53:25 -08:00