8c198e246d
## What changes were proposed in this pull request? This PR introduces the new SparkSession API for SparkR. `sparkR.session.getOrCreate()` and `sparkR.session.stop()` "getOrCreate" is a bit unusual in R but it's important to name this clearly. SparkR implementation should - SparkSession is the main entrypoint (vs SparkContext; due to limited functionality supported with SparkContext in SparkR) - SparkSession replaces SQLContext and HiveContext (both a wrapper around SparkSession, and because of API changes, supporting all 3 would be a lot more work) - Changes to SparkSession is mostly transparent to users due to SPARK-10903 - Full backward compatibility is expected - users should be able to initialize everything just in Spark 1.6.1 (`sparkR.init()`), but with deprecation warning - Mostly cosmetic changes to parameter list - users should be able to move to `sparkR.session.getOrCreate()` easily - An advanced syntax with named parameters (aka varargs aka "...") is supported; that should be closer to the Builder syntax that is in Scala/Python (which unfortunately does not work in R because it will look like this: `enableHiveSupport(config(config(master(appName(builder(), "foo"), "local"), "first", "value"), "next, "value"))` - Updating config on an existing SparkSession is supported, the behavior is the same as Python, in which config is applied to both SparkContext and SparkSession - Some SparkSession changes are not matched in SparkR, mostly because it would be breaking API change: `catalog` object, `createOrReplaceTempView` - Other SQLContext workarounds are replicated in SparkR, eg. `tables`, `tableNames` - `sparkR` shell is updated to use the SparkSession entrypoint (`sqlContext` is removed, just like with Scale/Python) - All tests are updated to use the SparkSession entrypoint - A bug in `read.jdbc` is fixed TODO - [x] Add more tests - [ ] Separate PR - update all roxygen2 doc coding example - [ ] Separate PR - update SparkR programming guide ## How was this patch tested? unit tests, manual tests shivaram sun-rui rxin Author: Felix Cheung <felixcheung_m@hotmail.com> Author: felixcheung <felixcheung_m@hotmail.com> Closes #13635 from felixcheung/rsparksession.
68 lines
3.2 KiB
R
68 lines
3.2 KiB
R
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
context("tests RDD function take()")
|
|
|
|
# Mock data
|
|
numVector <- c(-10:97)
|
|
numList <- list(sqrt(1), sqrt(2), sqrt(3), 4 ** 10)
|
|
strVector <- c("Dexter Morgan: I suppose I should be upset, even feel",
|
|
"violated, but I'm not. No, in fact, I think this is a friendly",
|
|
"message, like \"Hey, wanna play?\" and yes, I want to play. ",
|
|
"I really, really do.")
|
|
strList <- list("Dexter Morgan: Blood. Sometimes it sets my teeth on edge, ",
|
|
"other times it helps me control the chaos.",
|
|
"Dexter Morgan: Harry and Dorris Morgan did a wonderful job ",
|
|
"raising me. But they're both dead now. I didn't kill them. Honest.")
|
|
|
|
# JavaSparkContext handle
|
|
sparkSession <- sparkR.session()
|
|
sc <- callJStatic("org.apache.spark.sql.api.r.SQLUtils", "getJavaSparkContext", sparkSession)
|
|
|
|
test_that("take() gives back the original elements in correct count and order", {
|
|
numVectorRDD <- parallelize(sc, numVector, 10)
|
|
# case: number of elements to take is less than the size of the first partition
|
|
expect_equal(take(numVectorRDD, 1), as.list(head(numVector, n = 1)))
|
|
# case: number of elements to take is the same as the size of the first partition
|
|
expect_equal(take(numVectorRDD, 11), as.list(head(numVector, n = 11)))
|
|
# case: number of elements to take is greater than all elements
|
|
expect_equal(take(numVectorRDD, length(numVector)), as.list(numVector))
|
|
expect_equal(take(numVectorRDD, length(numVector) + 1), as.list(numVector))
|
|
|
|
numListRDD <- parallelize(sc, numList, 1)
|
|
numListRDD2 <- parallelize(sc, numList, 4)
|
|
expect_equal(take(numListRDD, 3), take(numListRDD2, 3))
|
|
expect_equal(take(numListRDD, 5), take(numListRDD2, 5))
|
|
expect_equal(take(numListRDD, 1), as.list(head(numList, n = 1)))
|
|
expect_equal(take(numListRDD2, 999), numList)
|
|
|
|
strVectorRDD <- parallelize(sc, strVector, 2)
|
|
strVectorRDD2 <- parallelize(sc, strVector, 3)
|
|
expect_equal(take(strVectorRDD, 4), as.list(strVector))
|
|
expect_equal(take(strVectorRDD2, 2), as.list(head(strVector, n = 2)))
|
|
|
|
strListRDD <- parallelize(sc, strList, 4)
|
|
strListRDD2 <- parallelize(sc, strList, 1)
|
|
expect_equal(take(strListRDD, 3), as.list(head(strList, n = 3)))
|
|
expect_equal(take(strListRDD2, 1), as.list(head(strList, n = 1)))
|
|
|
|
expect_equal(length(take(strListRDD, 0)), 0)
|
|
expect_equal(length(take(strVectorRDD, 0)), 0)
|
|
expect_equal(length(take(numListRDD, 0)), 0)
|
|
expect_equal(length(take(numVectorRDD, 0)), 0)
|
|
})
|