[MINOR][PYSPARK][DOCS] Fix examples in PySpark documentation
## What changes were proposed in this pull request? This PR proposes to fix wrongly indented examples in PySpark documentation ``` - >>> json_sdf = spark.readStream.format("json")\ - .schema(sdf_schema)\ - .load(tempfile.mkdtemp()) + >>> json_sdf = spark.readStream.format("json") \\ + ... .schema(sdf_schema) \\ + ... .load(tempfile.mkdtemp()) ``` ``` - people.filter(people.age > 30).join(department, people.deptId == department.id)\ + people.filter(people.age > 30).join(department, people.deptId == department.id) \\ ``` ``` - >>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, 1.23), (2, 4.56)])), \ - LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))] + >>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, 1.23), (2, 4.56)])), + ... LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))] ``` ``` - >>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, -1.23), (2, 4.56e-7)])), \ - LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))] + >>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, -1.23), (2, 4.56e-7)])), + ... LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))] ``` ``` - ... for x in iterator: - ... print(x) + ... for x in iterator: + ... print(x) ``` ## How was this patch tested? Manually tested. **Before** ![2016-09-26 8 36 02](https://cloud.githubusercontent.com/assets/6477701/18834471/05c7a478-8431-11e6-94bb-09aa37b12ddb.png) ![2016-09-26 9 22 16](https://cloud.githubusercontent.com/assets/6477701/18834472/06c8735c-8431-11e6-8775-78631eab0411.png) <img width="601" alt="2016-09-27 2 29 27" src="https://cloud.githubusercontent.com/assets/6477701/18861294/29c0d5b4-84bf-11e6-99c5-3c9d913c125d.png"> <img width="1056" alt="2016-09-27 2 29 58" src="https://cloud.githubusercontent.com/assets/6477701/18861298/31694cd8-84bf-11e6-9e61-9888cb8c2089.png"> <img width="1079" alt="2016-09-27 2 30 05" src="https://cloud.githubusercontent.com/assets/6477701/18861301/359722da-84bf-11e6-97f9-5f5365582d14.png"> **After** ![2016-09-26 9 29 47](https://cloud.githubusercontent.com/assets/6477701/18834467/0367f9da-8431-11e6-86d9-a490d3297339.png) ![2016-09-26 9 30 24](https://cloud.githubusercontent.com/assets/6477701/18834463/f870fae0-8430-11e6-9482-01fc47898492.png) <img width="515" alt="2016-09-27 2 28 19" src="https://cloud.githubusercontent.com/assets/6477701/18861305/3ff88b88-84bf-11e6-902c-9f725e8a8b10.png"> <img width="652" alt="2016-09-27 3 50 59" src="https://cloud.githubusercontent.com/assets/6477701/18863053/592fbc74-84ca-11e6-8dbf-99cf57947de8.png"> <img width="709" alt="2016-09-27 3 51 03" src="https://cloud.githubusercontent.com/assets/6477701/18863060/601607be-84ca-11e6-80aa-a401df41c321.png"> Author: hyukjinkwon <gurwls223@gmail.com> Closes #15242 from HyukjinKwon/minor-example-pyspark.
This commit is contained in:
parent
b2a7eedcdd
commit
2190037757
|
@ -140,8 +140,8 @@ class MLUtils(object):
|
|||
>>> from pyspark.mllib.regression import LabeledPoint
|
||||
>>> from glob import glob
|
||||
>>> from pyspark.mllib.util import MLUtils
|
||||
>>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, 1.23), (2, 4.56)])), \
|
||||
LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))]
|
||||
>>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, 1.23), (2, 4.56)])),
|
||||
... LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))]
|
||||
>>> tempFile = NamedTemporaryFile(delete=True)
|
||||
>>> tempFile.close()
|
||||
>>> MLUtils.saveAsLibSVMFile(sc.parallelize(examples), tempFile.name)
|
||||
|
@ -166,8 +166,8 @@ class MLUtils(object):
|
|||
>>> from tempfile import NamedTemporaryFile
|
||||
>>> from pyspark.mllib.util import MLUtils
|
||||
>>> from pyspark.mllib.regression import LabeledPoint
|
||||
>>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, -1.23), (2, 4.56e-7)])), \
|
||||
LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))]
|
||||
>>> examples = [LabeledPoint(1.1, Vectors.sparse(3, [(0, -1.23), (2, 4.56e-7)])),
|
||||
... LabeledPoint(0.0, Vectors.dense([1.01, 2.02, 3.03]))]
|
||||
>>> tempFile = NamedTemporaryFile(delete=True)
|
||||
>>> tempFile.close()
|
||||
>>> sc.parallelize(examples, 1).saveAsTextFile(tempFile.name)
|
||||
|
|
|
@ -754,8 +754,8 @@ class RDD(object):
|
|||
Applies a function to each partition of this RDD.
|
||||
|
||||
>>> def f(iterator):
|
||||
... for x in iterator:
|
||||
... print(x)
|
||||
... for x in iterator:
|
||||
... print(x)
|
||||
>>> sc.parallelize([1, 2, 3, 4, 5]).foreachPartition(f)
|
||||
"""
|
||||
def func(it):
|
||||
|
|
|
@ -61,7 +61,7 @@ class DataFrame(object):
|
|||
people = sqlContext.read.parquet("...")
|
||||
department = sqlContext.read.parquet("...")
|
||||
|
||||
people.filter(people.age > 30).join(department, people.deptId == department.id)\
|
||||
people.filter(people.age > 30).join(department, people.deptId == department.id) \\
|
||||
.groupBy(department.name, "gender").agg({"salary": "avg", "age": "max"})
|
||||
|
||||
.. versionadded:: 1.3
|
||||
|
|
|
@ -315,9 +315,9 @@ class DataStreamReader(OptionUtils):
|
|||
:param schema: optional :class:`pyspark.sql.types.StructType` for the input schema.
|
||||
:param options: all other string options
|
||||
|
||||
>>> json_sdf = spark.readStream.format("json")\
|
||||
.schema(sdf_schema)\
|
||||
.load(tempfile.mkdtemp())
|
||||
>>> json_sdf = spark.readStream.format("json") \\
|
||||
... .schema(sdf_schema) \\
|
||||
... .load(tempfile.mkdtemp())
|
||||
>>> json_sdf.isStreaming
|
||||
True
|
||||
>>> json_sdf.schema == sdf_schema
|
||||
|
|
Loading…
Reference in a new issue