spark-instrumented-optimizer/R/pkg/NAMESPACE
Yanbo Liang ba02f6cb5a [SPARK-12025][SPARKR] Rename some window rank function names for SparkR
Change ```cumeDist -> cume_dist, denseRank -> dense_rank, percentRank -> percent_rank, rowNumber -> row_number``` at SparkR side.
There are two reasons that we should make this change:
* We should follow the [naming convention rule of R](http://www.inside-r.org/node/230645)
* Spark DataFrame has deprecated the old convention (such as ```cumeDist```) and will remove it in Spark 2.0.

It's better to fix this issue before 1.6 release, otherwise we will make breaking API change.
cc shivaram sun-rui

Author: Yanbo Liang <ybliang8@gmail.com>

Closes #10016 from yanboliang/SPARK-12025.
2015-11-27 11:48:01 -08:00

279 lines
6.5 KiB
Plaintext

# Imports from base R
importFrom(methods, setGeneric, setMethod, setOldClass)
# Disable native libraries till we figure out how to package it
# See SPARKR-7839
#useDynLib(SparkR, stringHashCode)
# S3 methods exported
export("sparkR.init")
export("sparkR.stop")
export("print.jobj")
# MLlib integration
exportMethods("glm",
"predict",
"summary")
# Job group lifecycle management methods
export("setJobGroup",
"clearJobGroup",
"cancelJobGroup")
exportClasses("DataFrame")
exportMethods("arrange",
"as.data.frame",
"attach",
"cache",
"collect",
"coltypes",
"columns",
"count",
"cov",
"corr",
"crosstab",
"describe",
"dim",
"distinct",
"dropna",
"dtypes",
"except",
"explain",
"fillna",
"filter",
"first",
"freqItems",
"group_by",
"groupBy",
"head",
"insertInto",
"intersect",
"isLocal",
"join",
"limit",
"merge",
"mutate",
"na.omit",
"names",
"ncol",
"nrow",
"orderBy",
"persist",
"printSchema",
"rbind",
"registerTempTable",
"rename",
"repartition",
"sample",
"sample_frac",
"sampleBy",
"saveAsParquetFile",
"saveAsTable",
"saveDF",
"schema",
"select",
"selectExpr",
"show",
"showDF",
"subset",
"summarize",
"summary",
"take",
"transform",
"unionAll",
"unique",
"unpersist",
"where",
"with",
"withColumn",
"withColumnRenamed",
"write.df")
exportClasses("Column")
exportMethods("%in%",
"abs",
"acos",
"add_months",
"alias",
"approxCountDistinct",
"array_contains",
"asc",
"ascii",
"asin",
"atan",
"atan2",
"avg",
"base64",
"between",
"bin",
"bitwiseNOT",
"cast",
"cbrt",
"ceil",
"ceiling",
"column",
"concat",
"concat_ws",
"contains",
"conv",
"cos",
"cosh",
"count",
"countDistinct",
"crc32",
"cume_dist",
"date_add",
"date_format",
"date_sub",
"datediff",
"dayofmonth",
"dayofyear",
"dense_rank",
"desc",
"endsWith",
"exp",
"explode",
"expm1",
"expr",
"factorial",
"first",
"floor",
"format_number",
"format_string",
"from_unixtime",
"from_utc_timestamp",
"getField",
"getItem",
"greatest",
"hex",
"hour",
"hypot",
"ifelse",
"initcap",
"instr",
"isNaN",
"isNotNull",
"isNull",
"kurtosis",
"lag",
"last",
"last_day",
"lead",
"least",
"length",
"levenshtein",
"like",
"lit",
"locate",
"log",
"log10",
"log1p",
"log2",
"lower",
"lpad",
"ltrim",
"max",
"md5",
"mean",
"min",
"minute",
"month",
"months_between",
"n",
"n_distinct",
"nanvl",
"negate",
"next_day",
"ntile",
"otherwise",
"percent_rank",
"pmod",
"quarter",
"rand",
"randn",
"rank",
"regexp_extract",
"regexp_replace",
"reverse",
"rint",
"rlike",
"round",
"row_number",
"rpad",
"rtrim",
"second",
"sha1",
"sha2",
"shiftLeft",
"shiftRight",
"shiftRightUnsigned",
"sd",
"sign",
"signum",
"sin",
"sinh",
"size",
"skewness",
"sort_array",
"soundex",
"stddev",
"stddev_pop",
"stddev_samp",
"sqrt",
"startsWith",
"substr",
"substring_index",
"sum",
"sumDistinct",
"tan",
"tanh",
"toDegrees",
"toRadians",
"to_date",
"to_utc_timestamp",
"translate",
"trim",
"unbase64",
"unhex",
"unix_timestamp",
"upper",
"var",
"variance",
"var_pop",
"var_samp",
"weekofyear",
"when",
"year")
exportClasses("GroupedData")
exportMethods("agg")
export("sparkRSQL.init",
"sparkRHive.init")
export("as.DataFrame",
"cacheTable",
"clearCache",
"createDataFrame",
"createExternalTable",
"dropTempTable",
"jsonFile",
"loadDF",
"parquetFile",
"read.df",
"sql",
"table",
"tableNames",
"tables",
"uncacheTable")
export("structField",
"structField.jobj",
"structField.character",
"print.structField",
"structType",
"structType.jobj",
"structType.structField",
"print.structType")