spark-instrumented-optimizer/python/test_support
hyukjinkwon 7e5359be5c [SPARK-19610][SQL] Support parsing multiline CSV files
## What changes were proposed in this pull request?

This PR proposes the support for multiple lines for CSV by resembling the multiline supports in JSON datasource (in case of JSON, per file).

So, this PR introduces `wholeFile` option which makes the format not splittable and reads each whole file. Since Univocity parser can produces each row from a stream, it should be capable of parsing very large documents when the internal rows are fix in the memory.

## How was this patch tested?

Unit tests in `CSVSuite` and `tests.py`

Manual tests with a single 9GB CSV file in local file system, for example,

```scala
spark.read.option("wholeFile", true).option("inferSchema", true).csv("tmp.csv").count()
```

Author: hyukjinkwon <gurwls223@gmail.com>

Closes #16976 from HyukjinKwon/SPARK-19610.
2017-02-28 13:34:33 -08:00
..
hello [SPARK-17585][PYSPARK][CORE] PySpark SparkContext.addFile supports adding files recursively 2016-09-21 01:37:03 -07:00
sql [SPARK-19610][SQL] Support parsing multiline CSV files 2017-02-28 13:34:33 -08:00
SimpleHTTPServer.py [SPARK-3634] [PySpark] User's module should take precedence over system modules 2014-09-24 12:10:09 -07:00
userlib-0.1.zip [SPARK-4897] [PySpark] Python 3 support 2015-04-16 16:20:57 -07:00
userlibrary.py [SPARK-2627] [PySpark] have the build enforce PEP 8 automatically 2014-08-06 12:58:24 -07:00