214ba66a03
## What changes were proposed in this pull request? Add an install_spark function to the SparkR package. User can run `install_spark()` to install Spark to a local directory within R. Updates: Several changes have been made: - `install.spark()` - check existence of tar file in the cache folder, and download only if not found - trial priority of mirror_url look-up: user-provided -> preferred mirror site from apache website -> hardcoded backup option - use 2.0.0 - `sparkR.session()` - can install spark when not found in `SPARK_HOME` ## How was this patch tested? Manual tests, running the check-cran.sh script added in #14173. Author: Junyang Qian <junyangq@databricks.com> Closes #14258 from junyangq/SPARK-16579.
43 lines
797 B
Plaintext
43 lines
797 B
Plaintext
Package: SparkR
|
|
Type: Package
|
|
Title: R Frontend for Apache Spark
|
|
Version: 2.0.0
|
|
Date: 2016-07-07
|
|
Author: The Apache Software Foundation
|
|
Maintainer: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
|
|
Depends:
|
|
R (>= 3.0),
|
|
methods
|
|
Suggests:
|
|
testthat,
|
|
e1071,
|
|
survival
|
|
Description: The SparkR package provides an R frontend for Apache Spark.
|
|
License: Apache License (== 2.0)
|
|
Collate:
|
|
'schema.R'
|
|
'generics.R'
|
|
'jobj.R'
|
|
'column.R'
|
|
'group.R'
|
|
'RDD.R'
|
|
'pairRDD.R'
|
|
'DataFrame.R'
|
|
'SQLContext.R'
|
|
'WindowSpec.R'
|
|
'backend.R'
|
|
'broadcast.R'
|
|
'client.R'
|
|
'context.R'
|
|
'deserialize.R'
|
|
'functions.R'
|
|
'install.R'
|
|
'mllib.R'
|
|
'serialize.R'
|
|
'sparkR.R'
|
|
'stats.R'
|
|
'types.R'
|
|
'utils.R'
|
|
'window.R'
|
|
RoxygenNote: 5.0.1
|