f1d4e7e311
Author: Sun Rui <rui.sun@intel.com>
Closes #6183 from sun-rui/SPARK-7227 and squashes the following commits:
dd6f5b3 [Sun Rui] Rename readEnv() back to readMap(). Add alias na.omit() for dropna().
41cf725 [Sun Rui] [SPARK-7227][SPARKR] Support fillna / dropna in R DataFrame.
(cherry picked from commit 46576ab303
)
Signed-off-by: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
148 lines
3.3 KiB
Plaintext
148 lines
3.3 KiB
Plaintext
# Imports from base R
|
|
importFrom(methods, setGeneric, setMethod, setOldClass)
|
|
|
|
# Disable native libraries till we figure out how to package it
|
|
# See SPARKR-7839
|
|
#useDynLib(SparkR, stringHashCode)
|
|
|
|
# S3 methods exported
|
|
export("sparkR.init")
|
|
export("sparkR.stop")
|
|
export("print.jobj")
|
|
|
|
exportClasses("DataFrame")
|
|
|
|
exportMethods("arrange",
|
|
"cache",
|
|
"collect",
|
|
"columns",
|
|
"count",
|
|
"describe",
|
|
"distinct",
|
|
"dropna",
|
|
"dtypes",
|
|
"except",
|
|
"explain",
|
|
"fillna",
|
|
"filter",
|
|
"first",
|
|
"group_by",
|
|
"groupBy",
|
|
"head",
|
|
"insertInto",
|
|
"intersect",
|
|
"isLocal",
|
|
"join",
|
|
"limit",
|
|
"orderBy",
|
|
"mutate",
|
|
"names",
|
|
"persist",
|
|
"printSchema",
|
|
"registerTempTable",
|
|
"rename",
|
|
"repartition",
|
|
"sample",
|
|
"sample_frac",
|
|
"saveAsParquetFile",
|
|
"saveAsTable",
|
|
"saveDF",
|
|
"schema",
|
|
"select",
|
|
"selectExpr",
|
|
"show",
|
|
"showDF",
|
|
"summarize",
|
|
"take",
|
|
"unionAll",
|
|
"unpersist",
|
|
"where",
|
|
"withColumn",
|
|
"withColumnRenamed",
|
|
"write.df")
|
|
|
|
exportClasses("Column")
|
|
|
|
exportMethods("abs",
|
|
"acos",
|
|
"alias",
|
|
"approxCountDistinct",
|
|
"asc",
|
|
"asin",
|
|
"atan",
|
|
"atan2",
|
|
"avg",
|
|
"cast",
|
|
"cbrt",
|
|
"ceiling",
|
|
"contains",
|
|
"cos",
|
|
"cosh",
|
|
"countDistinct",
|
|
"desc",
|
|
"endsWith",
|
|
"exp",
|
|
"expm1",
|
|
"floor",
|
|
"getField",
|
|
"getItem",
|
|
"hypot",
|
|
"isNotNull",
|
|
"isNull",
|
|
"last",
|
|
"like",
|
|
"log",
|
|
"log10",
|
|
"log1p",
|
|
"lower",
|
|
"max",
|
|
"mean",
|
|
"min",
|
|
"n",
|
|
"n_distinct",
|
|
"rint",
|
|
"rlike",
|
|
"sign",
|
|
"sin",
|
|
"sinh",
|
|
"sqrt",
|
|
"startsWith",
|
|
"substr",
|
|
"sum",
|
|
"sumDistinct",
|
|
"tan",
|
|
"tanh",
|
|
"toDegrees",
|
|
"toRadians",
|
|
"upper")
|
|
|
|
exportClasses("GroupedData")
|
|
exportMethods("agg")
|
|
|
|
export("sparkRSQL.init",
|
|
"sparkRHive.init")
|
|
|
|
export("cacheTable",
|
|
"clearCache",
|
|
"createDataFrame",
|
|
"createExternalTable",
|
|
"dropTempTable",
|
|
"jsonFile",
|
|
"loadDF",
|
|
"parquetFile",
|
|
"read.df",
|
|
"sql",
|
|
"table",
|
|
"tableNames",
|
|
"tables",
|
|
"uncacheTable")
|
|
|
|
export("structField",
|
|
"structField.jobj",
|
|
"structField.character",
|
|
"print.structField",
|
|
"structType",
|
|
"structType.jobj",
|
|
"structType.structField",
|
|
"print.structType")
|