spark-instrumented-optimizer/python
Michael Armbrust 273c2fd08d [SQL] SPARK-1424 Generalize insertIntoTable functions on SchemaRDDs
This makes it possible to create tables and insert into them using the DSL and SQL for the scala and java apis.

Author: Michael Armbrust <michael@databricks.com>

Closes #354 from marmbrus/insertIntoTable and squashes the following commits:

6c6f227 [Michael Armbrust] Create random temporary files in python parquet unit tests.
f5e6d5c [Michael Armbrust] Merge remote-tracking branch 'origin/master' into insertIntoTable
765c506 [Michael Armbrust] Add to JavaAPI.
77b512c [Michael Armbrust] typos.
5c3ef95 [Michael Armbrust] use names for boolean args.
882afdf [Michael Armbrust] Change createTableAs to saveAsTable.  Clean up api annotations.
d07d94b [Michael Armbrust] Add tests, support for creating parquet files and hive tables.
fa3fe81 [Michael Armbrust] Make insertInto available on JavaSchemaRDD as well.  Add createTableAs function.
2014-04-15 20:40:40 -07:00
..
examples [WIP] SPARK-1430: Support sparse data in Python MLlib 2014-04-15 20:33:24 -07:00
lib Changes on top of Prashant's patch. 2014-01-03 18:30:17 -08:00
pyspark [SQL] SPARK-1424 Generalize insertIntoTable functions on SchemaRDDs 2014-04-15 20:40:40 -07:00
test_support License headers 2013-12-09 16:41:01 -08:00
.gitignore Rename top-level 'pyspark' directory to 'python' 2013-01-01 15:05:00 -08:00
epydoc.conf [WIP] SPARK-1430: Support sparse data in Python MLlib 2014-04-15 20:33:24 -07:00
run-tests [WIP] SPARK-1430: Support sparse data in Python MLlib 2014-04-15 20:33:24 -07:00