You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by sr...@apache.org on 2018/02/28 18:49:44 UTC

[38/51] [partial] spark-website git commit: Apache Spark 2.3.0 Release: Generated Docs

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/spark.survreg.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/spark.survreg.html b/site/docs/2.3.0/api/R/spark.survreg.html
new file mode 100644
index 0000000..e3e412d
--- /dev/null
+++ b/site/docs/2.3.0/api/R/spark.survreg.html
@@ -0,0 +1,156 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Accelerated Failure Time (AFT) Survival Regression Model</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for spark.survreg {SparkR}"><tr><td>spark.survreg {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Accelerated Failure Time (AFT) Survival Regression Model</h2>
+
+<h3>Description</h3>
+
+<p><code>spark.survreg</code> fits an accelerated failure time (AFT) survival regression model on
+a SparkDataFrame. Users can call <code>summary</code> to get a summary of the fitted AFT model,
+<code>predict</code> to make predictions on new data, and <code>write.ml</code>/<code>read.ml</code> to
+save/load fitted models.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+spark.survreg(data, formula, ...)
+
+## S4 method for signature 'SparkDataFrame,formula'
+spark.survreg(data, formula,
+  aggregationDepth = 2, stringIndexerOrderType = c("frequencyDesc",
+  "frequencyAsc", "alphabetDesc", "alphabetAsc"))
+
+## S4 method for signature 'AFTSurvivalRegressionModel'
+summary(object)
+
+## S4 method for signature 'AFTSurvivalRegressionModel'
+predict(object, newData)
+
+## S4 method for signature 'AFTSurvivalRegressionModel,character'
+write.ml(object, path,
+  overwrite = FALSE)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>data</code></td>
+<td>
+<p>a SparkDataFrame for training.</p>
+</td></tr>
+<tr valign="top"><td><code>formula</code></td>
+<td>
+<p>a symbolic description of the model to be fitted. Currently only a few formula
+operators are supported, including '~', ':', '+', and '-'.
+Note that operator '.' is not supported currently.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional arguments passed to the method.</p>
+</td></tr>
+<tr valign="top"><td><code>aggregationDepth</code></td>
+<td>
+<p>The depth for treeAggregate (greater than or equal to 2). If the
+dimensions of features or the number of partitions are large, this
+param could be adjusted to a larger size. This is an expert parameter.
+Default value should be good for most cases.</p>
+</td></tr>
+<tr valign="top"><td><code>stringIndexerOrderType</code></td>
+<td>
+<p>how to order categories of a string feature column. This is used to
+decide the base level of a string feature as the last category
+after ordering is dropped when encoding strings. Supported options
+are &quot;frequencyDesc&quot;, &quot;frequencyAsc&quot;, &quot;alphabetDesc&quot;, and
+&quot;alphabetAsc&quot;. The default value is &quot;frequencyDesc&quot;. When the
+ordering is set to &quot;alphabetDesc&quot;, this drops the same category
+as R when encoding strings.</p>
+</td></tr>
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a fitted AFT survival regression model.</p>
+</td></tr>
+<tr valign="top"><td><code>newData</code></td>
+<td>
+<p>a SparkDataFrame for testing.</p>
+</td></tr>
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>the directory where the model is saved.</p>
+</td></tr>
+<tr valign="top"><td><code>overwrite</code></td>
+<td>
+<p>overwrites or not if the output path already exists. Default is FALSE
+which means throw exception if the output path exists.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p><code>spark.survreg</code> returns a fitted AFT survival regression model.
+</p>
+<p><code>summary</code> returns summary information of the fitted model, which is a list.
+The list includes the model's <code>coefficients</code> (features, coefficients,
+intercept and log(scale)).
+</p>
+<p><code>predict</code> returns a SparkDataFrame containing predicted values
+on the original scale of the data (mean predicted value at scale = 1.0).
+</p>
+
+
+<h3>Note</h3>
+
+<p>spark.survreg since 2.0.0
+</p>
+<p>summary(AFTSurvivalRegressionModel) since 2.0.0
+</p>
+<p>predict(AFTSurvivalRegressionModel) since 2.0.0
+</p>
+<p>write.ml(AFTSurvivalRegressionModel, character) since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>survival: <a href="https://cran.r-project.org/package=survival">https://cran.r-project.org/package=survival</a>
+</p>
+<p><a href="write.ml.html">write.ml</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(ovarian)
+##D model &lt;- spark.survreg(df, Surv(futime, fustat) ~ ecog_ps + rx)
+##D 
+##D # get a summary of the model
+##D summary(model)
+##D 
+##D # make predictions
+##D predicted &lt;- predict(model, df)
+##D showDF(predicted)
+##D 
+##D # save and load the model
+##D path &lt;- &quot;path/to/model&quot;
+##D write.ml(model, path)
+##D savedModel &lt;- read.ml(path)
+##D summary(savedModel)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/spark.svmLinear.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/spark.svmLinear.html b/site/docs/2.3.0/api/R/spark.svmLinear.html
new file mode 100644
index 0000000..c81774f
--- /dev/null
+++ b/site/docs/2.3.0/api/R/spark.svmLinear.html
@@ -0,0 +1,176 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Linear SVM Model</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for spark.svmLinear {SparkR}"><tr><td>spark.svmLinear {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Linear SVM Model</h2>
+
+<h3>Description</h3>
+
+<p>Fits a linear SVM model against a SparkDataFrame, similar to svm in e1071 package.
+Currently only supports binary classification model with linear kernel.
+Users can print, make predictions on the produced model and save the model to the input path.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+spark.svmLinear(data, formula, ...)
+
+## S4 method for signature 'SparkDataFrame,formula'
+spark.svmLinear(data, formula,
+  regParam = 0, maxIter = 100, tol = 1e-06, standardization = TRUE,
+  threshold = 0, weightCol = NULL, aggregationDepth = 2,
+  handleInvalid = c("error", "keep", "skip"))
+
+## S4 method for signature 'LinearSVCModel'
+predict(object, newData)
+
+## S4 method for signature 'LinearSVCModel'
+summary(object)
+
+## S4 method for signature 'LinearSVCModel,character'
+write.ml(object, path, overwrite = FALSE)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>data</code></td>
+<td>
+<p>SparkDataFrame for training.</p>
+</td></tr>
+<tr valign="top"><td><code>formula</code></td>
+<td>
+<p>A symbolic description of the model to be fitted. Currently only a few formula
+operators are supported, including '~', '.', ':', '+', and '-'.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional arguments passed to the method.</p>
+</td></tr>
+<tr valign="top"><td><code>regParam</code></td>
+<td>
+<p>The regularization parameter. Only supports L2 regularization currently.</p>
+</td></tr>
+<tr valign="top"><td><code>maxIter</code></td>
+<td>
+<p>Maximum iteration number.</p>
+</td></tr>
+<tr valign="top"><td><code>tol</code></td>
+<td>
+<p>Convergence tolerance of iterations.</p>
+</td></tr>
+<tr valign="top"><td><code>standardization</code></td>
+<td>
+<p>Whether to standardize the training features before fitting the model.
+The coefficients of models will be always returned on the original scale,
+so it will be transparent for users. Note that with/without
+standardization, the models should be always converged to the same
+solution when no regularization is applied.</p>
+</td></tr>
+<tr valign="top"><td><code>threshold</code></td>
+<td>
+<p>The threshold in binary classification applied to the linear model prediction.
+This threshold can be any real number, where Inf will make all predictions 0.0
+and -Inf will make all predictions 1.0.</p>
+</td></tr>
+<tr valign="top"><td><code>weightCol</code></td>
+<td>
+<p>The weight column name.</p>
+</td></tr>
+<tr valign="top"><td><code>aggregationDepth</code></td>
+<td>
+<p>The depth for treeAggregate (greater than or equal to 2). If the
+dimensions of features or the number of partitions are large, this param
+could be adjusted to a larger size.
+This is an expert parameter. Default value should be good for most cases.</p>
+</td></tr>
+<tr valign="top"><td><code>handleInvalid</code></td>
+<td>
+<p>How to handle invalid data (unseen labels or NULL values) in features and
+label column of string type.
+Supported options: &quot;skip&quot; (filter out rows with invalid data),
+&quot;error&quot; (throw an error), &quot;keep&quot; (put invalid data in
+a special additional bucket, at index numLabels). Default
+is &quot;error&quot;.</p>
+</td></tr>
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a LinearSVCModel fitted by <code>spark.svmLinear</code>.</p>
+</td></tr>
+<tr valign="top"><td><code>newData</code></td>
+<td>
+<p>a SparkDataFrame for testing.</p>
+</td></tr>
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>The directory where the model is saved.</p>
+</td></tr>
+<tr valign="top"><td><code>overwrite</code></td>
+<td>
+<p>Overwrites or not if the output path already exists. Default is FALSE
+which means throw exception if the output path exists.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p><code>spark.svmLinear</code> returns a fitted linear SVM model.
+</p>
+<p><code>predict</code> returns the predicted values based on a LinearSVCModel.
+</p>
+<p><code>summary</code> returns summary information of the fitted model, which is a list.
+The list includes <code>coefficients</code> (coefficients of the fitted model),
+<code>numClasses</code> (number of classes), <code>numFeatures</code> (number of features).
+</p>
+
+
+<h3>Note</h3>
+
+<p>spark.svmLinear since 2.2.0
+</p>
+<p>predict(LinearSVCModel) since 2.2.0
+</p>
+<p>summary(LinearSVCModel) since 2.2.0
+</p>
+<p>write.ml(LogisticRegression, character) since 2.2.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D t &lt;- as.data.frame(Titanic)
+##D training &lt;- createDataFrame(t)
+##D model &lt;- spark.svmLinear(training, Survived ~ ., regParam = 0.5)
+##D summary &lt;- summary(model)
+##D 
+##D # fitted values on training data
+##D fitted &lt;- predict(model, training)
+##D 
+##D # save fitted model to input path
+##D path &lt;- &quot;path/to/model&quot;
+##D write.ml(model, path)
+##D 
+##D # can also read back the saved model and predict
+##D # Note that summary deos not work on loaded model
+##D savedModel &lt;- read.ml(path)
+##D summary(savedModel)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.callJMethod.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.callJMethod.html b/site/docs/2.3.0/api/R/sparkR.callJMethod.html
new file mode 100644
index 0000000..1f7cb04
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.callJMethod.html
@@ -0,0 +1,90 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Call Java Methods</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.callJMethod {SparkR}"><tr><td>sparkR.callJMethod {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Call Java Methods</h2>
+
+<h3>Description</h3>
+
+<p>Call a Java method in the JVM running the Spark driver. The return
+values are automatically converted to R objects for simple objects. Other
+values are returned as &quot;jobj&quot; which are references to objects on JVM.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.callJMethod(x, methodName, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>object to invoke the method on. Should be a &quot;jobj&quot; created by newJObject.</p>
+</td></tr>
+<tr valign="top"><td><code>methodName</code></td>
+<td>
+<p>method name to call.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>parameters to pass to the Java method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is a low level function to access the JVM directly and should only be used
+for advanced use cases. The arguments and return values that are primitive R
+types (like integer, numeric, character, lists) are automatically translated to/from
+Java types (like Integer, Double, String, Array). A full list can be found in
+serialize.R and deserialize.R in the Apache Spark code base.
+</p>
+
+
+<h3>Value</h3>
+
+<p>the return value of the Java method. Either returned as a R object
+if it can be deserialized or returned as a &quot;jobj&quot;. See details section for more.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.callJMethod since 2.0.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.callJStatic.html">sparkR.callJStatic</a>, <a href="sparkR.newJObject.html">sparkR.newJObject</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session() # Need to have a Spark JVM running before calling newJObject
+##D # Create a Java ArrayList and populate it
+##D jarray &lt;- sparkR.newJObject(&quot;java.util.ArrayList&quot;)
+##D sparkR.callJMethod(jarray, &quot;add&quot;, 42L)
+##D sparkR.callJMethod(jarray, &quot;get&quot;, 0L) # Will print 42
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.callJStatic.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.callJStatic.html b/site/docs/2.3.0/api/R/sparkR.callJStatic.html
new file mode 100644
index 0000000..036ff08
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.callJStatic.html
@@ -0,0 +1,88 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Call Static Java Methods</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.callJStatic {SparkR}"><tr><td>sparkR.callJStatic {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Call Static Java Methods</h2>
+
+<h3>Description</h3>
+
+<p>Call a static method in the JVM running the Spark driver. The return
+value is automatically converted to R objects for simple objects. Other
+values are returned as &quot;jobj&quot; which are references to objects on JVM.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.callJStatic(x, methodName, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>fully qualified Java class name that contains the static method to invoke.</p>
+</td></tr>
+<tr valign="top"><td><code>methodName</code></td>
+<td>
+<p>name of static method to invoke.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>parameters to pass to the Java method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is a low level function to access the JVM directly and should only be used
+for advanced use cases. The arguments and return values that are primitive R
+types (like integer, numeric, character, lists) are automatically translated to/from
+Java types (like Integer, Double, String, Array). A full list can be found in
+serialize.R and deserialize.R in the Apache Spark code base.
+</p>
+
+
+<h3>Value</h3>
+
+<p>the return value of the Java method. Either returned as a R object
+if it can be deserialized or returned as a &quot;jobj&quot;. See details section for more.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.callJStatic since 2.0.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.callJMethod.html">sparkR.callJMethod</a>, <a href="sparkR.newJObject.html">sparkR.newJObject</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session() # Need to have a Spark JVM running before calling callJStatic
+##D sparkR.callJStatic(&quot;java.lang.System&quot;, &quot;currentTimeMillis&quot;)
+##D sparkR.callJStatic(&quot;java.lang.System&quot;, &quot;getProperty&quot;, &quot;java.home&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.conf.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.conf.html b/site/docs/2.3.0/api/R/sparkR.conf.html
new file mode 100644
index 0000000..e297da0
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.conf.html
@@ -0,0 +1,68 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Get Runtime Config from the current active SparkSession</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.conf {SparkR}"><tr><td>sparkR.conf {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Get Runtime Config from the current active SparkSession</h2>
+
+<h3>Description</h3>
+
+<p>Get Runtime Config from the current active SparkSession.
+To change SparkSession Runtime Config, please see <code>sparkR.session()</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.conf(key, defaultValue)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>key</code></td>
+<td>
+<p>(optional) The key of the config to get, if omitted, all config is returned</p>
+</td></tr>
+<tr valign="top"><td><code>defaultValue</code></td>
+<td>
+<p>(optional) The default value of the config to return if they config is not
+set, if omitted, the call fails if the config key is not set</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>a list of config values with keys as their names
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.conf since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D allConfigs &lt;- sparkR.conf()
+##D masterValue &lt;- unlist(sparkR.conf(&quot;spark.master&quot;))
+##D namedConfig &lt;- sparkR.conf(&quot;spark.executor.memory&quot;, &quot;0g&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.init-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.init-deprecated.html b/site/docs/2.3.0/api/R/sparkR.init-deprecated.html
new file mode 100644
index 0000000..bdec09d
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.init-deprecated.html
@@ -0,0 +1,92 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: (Deprecated) Initialize a new Spark Context</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.init {SparkR}"><tr><td>sparkR.init {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Initialize a new Spark Context</h2>
+
+<h3>Description</h3>
+
+<p>This function initializes a new SparkContext.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.init(master = "", appName = "SparkR",
+  sparkHome = Sys.getenv("SPARK_HOME"), sparkEnvir = list(),
+  sparkExecutorEnv = list(), sparkJars = "", sparkPackages = "")
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>master</code></td>
+<td>
+<p>The Spark master URL</p>
+</td></tr>
+<tr valign="top"><td><code>appName</code></td>
+<td>
+<p>Application name to register with cluster manager</p>
+</td></tr>
+<tr valign="top"><td><code>sparkHome</code></td>
+<td>
+<p>Spark Home directory</p>
+</td></tr>
+<tr valign="top"><td><code>sparkEnvir</code></td>
+<td>
+<p>Named list of environment variables to set on worker nodes</p>
+</td></tr>
+<tr valign="top"><td><code>sparkExecutorEnv</code></td>
+<td>
+<p>Named list of environment variables to be used when launching executors</p>
+</td></tr>
+<tr valign="top"><td><code>sparkJars</code></td>
+<td>
+<p>Character vector of jar files to pass to the worker nodes</p>
+</td></tr>
+<tr valign="top"><td><code>sparkPackages</code></td>
+<td>
+<p>Character vector of package coordinates</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>sparkR.init since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.session.html">sparkR.session</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sc &lt;- sparkR.init(&quot;local[2]&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;)
+##D sc &lt;- sparkR.init(&quot;local[2]&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;,
+##D                  list(spark.executor.memory=&quot;1g&quot;))
+##D sc &lt;- sparkR.init(&quot;yarn-client&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;,
+##D                  list(spark.executor.memory=&quot;4g&quot;),
+##D                  list(LD_LIBRARY_PATH=&quot;/directory of JVM libraries (libjvm.so) on workers/&quot;),
+##D                  c(&quot;one.jar&quot;, &quot;two.jar&quot;, &quot;three.jar&quot;),
+##D                  c(&quot;com.databricks:spark-avro_2.11:2.0.1&quot;))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.newJObject.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.newJObject.html b/site/docs/2.3.0/api/R/sparkR.newJObject.html
new file mode 100644
index 0000000..1e7e38f
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.newJObject.html
@@ -0,0 +1,86 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create Java Objects</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.newJObject {SparkR}"><tr><td>sparkR.newJObject {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create Java Objects</h2>
+
+<h3>Description</h3>
+
+<p>Create a new Java object in the JVM running the Spark driver. The return
+value is automatically converted to an R object for simple objects. Other
+values are returned as a &quot;jobj&quot; which is a reference to an object on JVM.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.newJObject(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>fully qualified Java class name.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>arguments to be passed to the constructor.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is a low level function to access the JVM directly and should only be used
+for advanced use cases. The arguments and return values that are primitive R
+types (like integer, numeric, character, lists) are automatically translated to/from
+Java types (like Integer, Double, String, Array). A full list can be found in
+serialize.R and deserialize.R in the Apache Spark code base.
+</p>
+
+
+<h3>Value</h3>
+
+<p>the object created. Either returned as a R object
+if it can be deserialized or returned as a &quot;jobj&quot;. See details section for more.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.newJObject since 2.0.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.callJMethod.html">sparkR.callJMethod</a>, <a href="sparkR.callJStatic.html">sparkR.callJStatic</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session() # Need to have a Spark JVM running before calling newJObject
+##D # Create a Java ArrayList and populate it
+##D jarray &lt;- sparkR.newJObject(&quot;java.util.ArrayList&quot;)
+##D sparkR.callJMethod(jarray, &quot;add&quot;, 42L)
+##D sparkR.callJMethod(jarray, &quot;get&quot;, 0L) # Will print 42
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.session.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.session.html b/site/docs/2.3.0/api/R/sparkR.session.html
new file mode 100644
index 0000000..c002136
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.session.html
@@ -0,0 +1,114 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Get the existing SparkSession or initialize a new...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.session {SparkR}"><tr><td>sparkR.session {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Get the existing SparkSession or initialize a new SparkSession.</h2>
+
+<h3>Description</h3>
+
+<p>SparkSession is the entry point into SparkR. <code>sparkR.session</code> gets the existing
+SparkSession or initializes a new SparkSession.
+Additional Spark properties can be set in <code>...</code>, and these named parameters take priority
+over values in <code>master</code>, <code>appName</code>, named lists of <code>sparkConfig</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.session(master = "", appName = "SparkR",
+  sparkHome = Sys.getenv("SPARK_HOME"), sparkConfig = list(),
+  sparkJars = "", sparkPackages = "", enableHiveSupport = TRUE, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>master</code></td>
+<td>
+<p>the Spark master URL.</p>
+</td></tr>
+<tr valign="top"><td><code>appName</code></td>
+<td>
+<p>application name to register with cluster manager.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkHome</code></td>
+<td>
+<p>Spark Home directory.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkConfig</code></td>
+<td>
+<p>named list of Spark configuration to set on worker nodes.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkJars</code></td>
+<td>
+<p>character vector of jar files to pass to the worker nodes.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkPackages</code></td>
+<td>
+<p>character vector of package coordinates</p>
+</td></tr>
+<tr valign="top"><td><code>enableHiveSupport</code></td>
+<td>
+<p>enable support for Hive, fallback if not built with Hive support; once
+set, this cannot be turned off on an existing session</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>named Spark properties passed to the method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>When called in an interactive session, this method checks for the Spark installation, and, if not
+found, it will be downloaded and cached automatically. Alternatively, <code>install.spark</code> can
+be called manually.
+</p>
+<p>A default warehouse is created automatically in the current directory when a managed table is
+created via <code>sql</code> statement <code>CREATE TABLE</code>, for example. To change the location of the
+warehouse, set the named parameter <code>spark.sql.warehouse.dir</code> to the SparkSession. Along with
+the warehouse, an accompanied metastore may also be automatically created in the current
+directory when a new SparkSession is initialized with <code>enableHiveSupport</code> set to
+<code>TRUE</code>, which is the default. For more details, refer to Hive configuration at
+<a href="http://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables">http://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables</a>.
+</p>
+<p>For details on how to initialize and use SparkR, refer to SparkR programming guide at
+<a href="http://spark.apache.org/docs/latest/sparkr.html#starting-up-sparksession">http://spark.apache.org/docs/latest/sparkr.html#starting-up-sparksession</a>.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.session since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df &lt;- read.json(path)
+##D 
+##D sparkR.session(&quot;local[2]&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;)
+##D sparkR.session(&quot;yarn-client&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;,
+##D                list(spark.executor.memory=&quot;4g&quot;),
+##D                c(&quot;one.jar&quot;, &quot;two.jar&quot;, &quot;three.jar&quot;),
+##D                c(&quot;com.databricks:spark-avro_2.11:2.0.1&quot;))
+##D sparkR.session(spark.master = &quot;yarn-client&quot;, spark.executor.memory = &quot;4g&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.session.stop.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.session.stop.html b/site/docs/2.3.0/api/R/sparkR.session.stop.html
new file mode 100644
index 0000000..fe4cb94
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.session.stop.html
@@ -0,0 +1,39 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Stop the Spark Session and Spark Context</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for sparkR.session.stop {SparkR}"><tr><td>sparkR.session.stop {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Stop the Spark Session and Spark Context</h2>
+
+<h3>Description</h3>
+
+<p>Stop the Spark Session and Spark Context.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.session.stop()
+
+sparkR.stop()
+</pre>
+
+
+<h3>Details</h3>
+
+<p>Also terminates the backend this R session is connected to.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.session.stop since 2.0.0
+</p>
+<p>sparkR.stop since 1.4.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.uiWebUrl.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.uiWebUrl.html b/site/docs/2.3.0/api/R/sparkR.uiWebUrl.html
new file mode 100644
index 0000000..e316dd9
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.uiWebUrl.html
@@ -0,0 +1,50 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Get the URL of the SparkUI instance for the current active...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.uiWebUrl {SparkR}"><tr><td>sparkR.uiWebUrl {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Get the URL of the SparkUI instance for the current active SparkSession</h2>
+
+<h3>Description</h3>
+
+<p>Get the URL of the SparkUI instance for the current active SparkSession.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.uiWebUrl()
+</pre>
+
+
+<h3>Value</h3>
+
+<p>the SparkUI URL, or NA if it is disabled, or not started.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.uiWebUrl since 2.1.1
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D url &lt;- sparkR.uiWebUrl()
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkR.version.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkR.version.html b/site/docs/2.3.0/api/R/sparkR.version.html
new file mode 100644
index 0000000..22ee2f3
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkR.version.html
@@ -0,0 +1,50 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Get version of Spark on which this application is running</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.version {SparkR}"><tr><td>sparkR.version {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Get version of Spark on which this application is running</h2>
+
+<h3>Description</h3>
+
+<p>Get version of Spark on which this application is running.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.version()
+</pre>
+
+
+<h3>Value</h3>
+
+<p>a character string of the Spark version
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.version since 2.0.1
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D version &lt;- sparkR.version()
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkRHive.init-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkRHive.init-deprecated.html b/site/docs/2.3.0/api/R/sparkRHive.init-deprecated.html
new file mode 100644
index 0000000..b5df74e
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkRHive.init-deprecated.html
@@ -0,0 +1,67 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: (Deprecated) Initialize a new HiveContext</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkRHive.init {SparkR}"><tr><td>sparkRHive.init {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Initialize a new HiveContext</h2>
+
+<h3>Description</h3>
+
+<p>This function creates a HiveContext from an existing JavaSparkContext
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkRHive.init(jsc = NULL)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jsc</code></td>
+<td>
+<p>The existing JavaSparkContext created with SparkR.init()</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Starting SparkR 2.0, a SparkSession is initialized and returned instead.
+This API is deprecated and kept for backward compatibility only.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkRHive.init since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.session.html">sparkR.session</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sc &lt;- sparkR.init()
+##D sqlContext &lt;- sparkRHive.init(sc)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sparkRSQL.init-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sparkRSQL.init-deprecated.html b/site/docs/2.3.0/api/R/sparkRSQL.init-deprecated.html
new file mode 100644
index 0000000..dd93e8f
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sparkRSQL.init-deprecated.html
@@ -0,0 +1,68 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: (Deprecated) Initialize a new SQLContext</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkRSQL.init {SparkR}"><tr><td>sparkRSQL.init {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Initialize a new SQLContext</h2>
+
+<h3>Description</h3>
+
+<p>This function creates a SparkContext from an existing JavaSparkContext and
+then uses it to initialize a new SQLContext
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkRSQL.init(jsc = NULL)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jsc</code></td>
+<td>
+<p>The existing JavaSparkContext created with SparkR.init()</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Starting SparkR 2.0, a SparkSession is initialized and returned instead.
+This API is deprecated and kept for backward compatibility only.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkRSQL.init since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.session.html">sparkR.session</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sc &lt;- sparkR.init()
+##D sqlContext &lt;- sparkRSQL.init(sc)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/sql.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/sql.html b/site/docs/2.3.0/api/R/sql.html
new file mode 100644
index 0000000..cebd22e
--- /dev/null
+++ b/site/docs/2.3.0/api/R/sql.html
@@ -0,0 +1,64 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: SQL Query</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sql {SparkR}"><tr><td>sql {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>SQL Query</h2>
+
+<h3>Description</h3>
+
+<p>Executes a SQL query using Spark, returning the result as a SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+sql(sqlQuery)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>sqlQuery</code></td>
+<td>
+<p>A character vector containing the SQL query</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>sql since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D createOrReplaceTempView(df, &quot;table&quot;)
+##D new_df &lt;- sql(&quot;SELECT * FROM table&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/startsWith.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/startsWith.html b/site/docs/2.3.0/api/R/startsWith.html
new file mode 100644
index 0000000..ebff84f
--- /dev/null
+++ b/site/docs/2.3.0/api/R/startsWith.html
@@ -0,0 +1,56 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: startsWith</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for startsWith {SparkR}"><tr><td>startsWith {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>startsWith</h2>
+
+<h3>Description</h3>
+
+<p>Determines if entries of x start with string (entries of) prefix respectively,
+where strings are recycled to common lengths.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+startsWith(x, prefix)
+
+## S4 method for signature 'Column'
+startsWith(x, prefix)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>vector of character string whose &quot;starts&quot; are considered</p>
+</td></tr>
+<tr valign="top"><td><code>prefix</code></td>
+<td>
+<p>character vector (often of length one)</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>startsWith since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="alias.html">alias</a></code>,
+<code><a href="between.html">between</a></code>, <code><a href="cast.html">cast</a></code>,
+<code><a href="endsWith.html">endsWith</a></code>, <code><a href="otherwise.html">otherwise</a></code>,
+<code><a href="over.html">over</a></code>, <code><a href="substr.html">substr</a></code>
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/status.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/status.html b/site/docs/2.3.0/api/R/status.html
new file mode 100644
index 0000000..0004325
--- /dev/null
+++ b/site/docs/2.3.0/api/R/status.html
@@ -0,0 +1,65 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: status</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for status {SparkR}"><tr><td>status {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>status</h2>
+
+<h3>Description</h3>
+
+<p>Prints the current status of the query in JSON format.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+status(x)
+
+## S4 method for signature 'StreamingQuery'
+status(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a StreamingQuery.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>status(StreamingQuery) since 2.2.0
+</p>
+<p>experimental
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other StreamingQuery methods: <code><a href="awaitTermination.html">awaitTermination</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="isActive.html">isActive</a></code>,
+<code><a href="lastProgress.html">lastProgress</a></code>, <code><a href="queryName.html">queryName</a></code>,
+<code><a href="stopQuery.html">stopQuery</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run:  status(sq) 
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/stopQuery.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/stopQuery.html b/site/docs/2.3.0/api/R/stopQuery.html
new file mode 100644
index 0000000..87a9792
--- /dev/null
+++ b/site/docs/2.3.0/api/R/stopQuery.html
@@ -0,0 +1,66 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: stopQuery</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for stopQuery {SparkR}"><tr><td>stopQuery {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>stopQuery</h2>
+
+<h3>Description</h3>
+
+<p>Stops the execution of this query if it is running. This method blocks until the execution is
+stopped.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+stopQuery(x)
+
+## S4 method for signature 'StreamingQuery'
+stopQuery(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a StreamingQuery.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>stopQuery(StreamingQuery) since 2.2.0
+</p>
+<p>experimental
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other StreamingQuery methods: <code><a href="awaitTermination.html">awaitTermination</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="isActive.html">isActive</a></code>,
+<code><a href="lastProgress.html">lastProgress</a></code>, <code><a href="queryName.html">queryName</a></code>,
+<code><a href="status.html">status</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run:  stopQuery(sq) 
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/storageLevel.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/storageLevel.html b/site/docs/2.3.0/api/R/storageLevel.html
new file mode 100644
index 0000000..20b148d
--- /dev/null
+++ b/site/docs/2.3.0/api/R/storageLevel.html
@@ -0,0 +1,105 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: StorageLevel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for storageLevel {SparkR}"><tr><td>storageLevel {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>StorageLevel</h2>
+
+<h3>Description</h3>
+
+<p>Get storagelevel of this SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'SparkDataFrame'
+storageLevel(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>the SparkDataFrame to get the storageLevel.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>storageLevel since 2.1.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="except.html">except</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="filter.html">filter</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapply.html">gapply</a></code>, <code><a href="getNumPartitions.html">getNumPartitions</a></code>,
+<code><a href="groupBy.html">group_by</a></code>, <code><a href="head.html">head</a></code>,
+<code><a href="hint.html">hint</a></code>, <code><a href="histogram.html">histogram</a></code>,
+<code><a href="insertInto.html">insertInto</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartition.html">repartition</a></code>,
+<code><a href="rollup.html">rollup</a></code>, <code><a href="sample.html">sample</a></code>,
+<code><a href="saveAsTable.html">saveAsTable</a></code>, <code><a href="schema.html">schema</a></code>,
+<code><a href="selectExpr.html">selectExpr</a></code>, <code><a href="select.html">select</a></code>,
+<code><a href="showDF.html">showDF</a></code>, <code><a href="show.html">show</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D persist(df, &quot;MEMORY_AND_DISK&quot;)
+##D storageLevel(df)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/str.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/str.html b/site/docs/2.3.0/api/R/str.html
new file mode 100644
index 0000000..3762759
--- /dev/null
+++ b/site/docs/2.3.0/api/R/str.html
@@ -0,0 +1,106 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Compactly display the structure of a dataset</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for str {SparkR}"><tr><td>str {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Compactly display the structure of a dataset</h2>
+
+<h3>Description</h3>
+
+<p>Display the structure of a SparkDataFrame, including column names, column types, as well as a
+a small sample of rows.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'SparkDataFrame'
+str(object)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a SparkDataFrame</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>str since 1.6.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="except.html">except</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="filter.html">filter</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapply.html">gapply</a></code>, <code><a href="getNumPartitions.html">getNumPartitions</a></code>,
+<code><a href="groupBy.html">group_by</a></code>, <code><a href="head.html">head</a></code>,
+<code><a href="hint.html">hint</a></code>, <code><a href="histogram.html">histogram</a></code>,
+<code><a href="insertInto.html">insertInto</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartition.html">repartition</a></code>,
+<code><a href="rollup.html">rollup</a></code>, <code><a href="sample.html">sample</a></code>,
+<code><a href="saveAsTable.html">saveAsTable</a></code>, <code><a href="schema.html">schema</a></code>,
+<code><a href="selectExpr.html">selectExpr</a></code>, <code><a href="select.html">select</a></code>,
+<code><a href="showDF.html">showDF</a></code>, <code><a href="show.html">show</a></code>,
+<code><a href="storageLevel.html">storageLevel</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Create a SparkDataFrame from the Iris dataset
+##D irisDF &lt;- createDataFrame(iris)
+##D 
+##D # Show the structure of the SparkDataFrame
+##D str(irisDF)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/structField.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/structField.html b/site/docs/2.3.0/api/R/structField.html
new file mode 100644
index 0000000..4b57ac8
--- /dev/null
+++ b/site/docs/2.3.0/api/R/structField.html
@@ -0,0 +1,83 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: structField</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for structField {SparkR}"><tr><td>structField {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>structField</h2>
+
+<h3>Description</h3>
+
+<p>Create a structField object that contains the metadata for a single field in a schema.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+structField(x, ...)
+
+## S3 method for class 'jobj'
+structField(x, ...)
+
+## S3 method for class 'character'
+structField(x, type, nullable = TRUE, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>the name of the field.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s) passed to the method.</p>
+</td></tr>
+<tr valign="top"><td><code>type</code></td>
+<td>
+<p>The data type of the field</p>
+</td></tr>
+<tr valign="top"><td><code>nullable</code></td>
+<td>
+<p>A logical vector indicating whether or not the field is nullable</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A structField object.
+</p>
+
+
+<h3>Note</h3>
+
+<p>structField since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D field1 &lt;- structField(&quot;a&quot;, &quot;integer&quot;)
+##D field2 &lt;- structField(&quot;c&quot;, &quot;string&quot;)
+##D field3 &lt;- structField(&quot;avg&quot;, &quot;double&quot;)
+##D schema &lt;- structType(field1, field2, field3)
+##D df1 &lt;- gapply(df, list(&quot;a&quot;, &quot;c&quot;),
+##D               function(key, x) { y &lt;- data.frame(key, mean(x$b), stringsAsFactors = FALSE) },
+##D               schema)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/structType.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/structType.html b/site/docs/2.3.0/api/R/structType.html
new file mode 100644
index 0000000..9a1bbc1
--- /dev/null
+++ b/site/docs/2.3.0/api/R/structType.html
@@ -0,0 +1,83 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: structType</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for structType {SparkR}"><tr><td>structType {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>structType</h2>
+
+<h3>Description</h3>
+
+<p>Create a structType object that contains the metadata for a SparkDataFrame. Intended for
+use with createDataFrame and toDF.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+structType(x, ...)
+
+## S3 method for class 'jobj'
+structType(x, ...)
+
+## S3 method for class 'structField'
+structType(x, ...)
+
+## S3 method for class 'character'
+structType(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a structField object (created with the <code>structField</code> method). Since Spark 2.3,
+this can be a DDL-formatted string, which is a comma separated list of field
+definitions, e.g., &quot;a INT, b STRING&quot;.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional structField objects</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>a structType object
+</p>
+
+
+<h3>Note</h3>
+
+<p>structType since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D schema &lt;- structType(structField(&quot;a&quot;, &quot;integer&quot;), structField(&quot;c&quot;, &quot;string&quot;),
+##D                       structField(&quot;avg&quot;, &quot;double&quot;))
+##D df1 &lt;- gapply(df, list(&quot;a&quot;, &quot;c&quot;),
+##D               function(key, x) { y &lt;- data.frame(key, mean(x$b), stringsAsFactors = FALSE) },
+##D               schema)
+##D schema &lt;- structType(&quot;a INT, c STRING, avg DOUBLE&quot;)
+##D df1 &lt;- gapply(df, list(&quot;a&quot;, &quot;c&quot;),
+##D               function(key, x) { y &lt;- data.frame(key, mean(x$b), stringsAsFactors = FALSE) },
+##D               schema)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/subset.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/subset.html b/site/docs/2.3.0/api/R/subset.html
new file mode 100644
index 0000000..4c81c49
--- /dev/null
+++ b/site/docs/2.3.0/api/R/subset.html
@@ -0,0 +1,170 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Subset</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for subset {SparkR}"><tr><td>subset {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Subset</h2>
+
+<h3>Description</h3>
+
+<p>Return subsets of SparkDataFrame according to given conditions
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+subset(x, ...)
+
+## S4 method for signature 'SparkDataFrame,numericOrcharacter'
+x[[i]]
+
+## S4 replacement method for signature 'SparkDataFrame,numericOrcharacter'
+x[[i]] &lt;- value
+
+## S4 method for signature 'SparkDataFrame'
+x[i, j, ..., drop = F]
+
+## S4 method for signature 'SparkDataFrame'
+subset(x, subset, select, drop = F, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>currently not used.</p>
+</td></tr>
+<tr valign="top"><td><code>i, subset</code></td>
+<td>
+<p>(Optional) a logical expression to filter on rows.
+For extract operator [[ and replacement operator [[&lt;-, the indexing parameter for
+a single Column.</p>
+</td></tr>
+<tr valign="top"><td><code>value</code></td>
+<td>
+<p>a Column or an atomic vector in the length of 1 as literal value, or <code>NULL</code>.
+If <code>NULL</code>, the specified Column is dropped.</p>
+</td></tr>
+<tr valign="top"><td><code>j, select</code></td>
+<td>
+<p>expression for the single Column or a list of columns to select from the
+SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>drop</code></td>
+<td>
+<p>if TRUE, a Column will be returned if the resulting dataset has only one column.
+Otherwise, a SparkDataFrame will always be returned.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A new SparkDataFrame containing only the rows that meet the condition with selected
+columns.
+</p>
+
+
+<h3>Note</h3>
+
+<p>[[ since 1.4.0
+</p>
+<p>[[&lt;- since 2.1.1
+</p>
+<p>[ since 1.4.0
+</p>
+<p>subset since 1.5.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="withColumn.html">withColumn</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="except.html">except</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="filter.html">filter</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapply.html">gapply</a></code>, <code><a href="getNumPartitions.html">getNumPartitions</a></code>,
+<code><a href="groupBy.html">group_by</a></code>, <code><a href="head.html">head</a></code>,
+<code><a href="hint.html">hint</a></code>, <code><a href="histogram.html">histogram</a></code>,
+<code><a href="insertInto.html">insertInto</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartition.html">repartition</a></code>,
+<code><a href="rollup.html">rollup</a></code>, <code><a href="sample.html">sample</a></code>,
+<code><a href="saveAsTable.html">saveAsTable</a></code>, <code><a href="schema.html">schema</a></code>,
+<code><a href="selectExpr.html">selectExpr</a></code>, <code><a href="select.html">select</a></code>,
+<code><a href="showDF.html">showDF</a></code>, <code><a href="show.html">show</a></code>,
+<code><a href="storageLevel.html">storageLevel</a></code>, <code><a href="str.html">str</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+<p>Other subsetting functions: <code><a href="filter.html">filter</a></code>,
+<code><a href="select.html">select</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D   # Columns can be selected using [[ and [
+##D   df[[2]] == df[[&quot;age&quot;]]
+##D   df[,2] == df[,&quot;age&quot;]
+##D   df[,c(&quot;name&quot;, &quot;age&quot;)]
+##D   # Or to filter rows
+##D   df[df$age &gt; 20,]
+##D   # SparkDataFrame can be subset on both rows and Columns
+##D   df[df$name == &quot;Smith&quot;, c(1,2)]
+##D   df[df$age %in% c(19, 30), 1:2]
+##D   subset(df, df$age %in% c(19, 30), 1:2)
+##D   subset(df, df$age %in% c(19), select = c(1,2))
+##D   subset(df, select = c(1,2))
+##D   # Columns can be selected and set
+##D   df[[&quot;age&quot;]] &lt;- 23
+##D   df[[1]] &lt;- df$age
+##D   df[[2]] &lt;- NULL # drop column
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/substr.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/substr.html b/site/docs/2.3.0/api/R/substr.html
new file mode 100644
index 0000000..3a3b55f
--- /dev/null
+++ b/site/docs/2.3.0/api/R/substr.html
@@ -0,0 +1,57 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: substr</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for substr {SparkR}"><tr><td>substr {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>substr</h2>
+
+<h3>Description</h3>
+
+<p>An expression that returns a substring.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'Column'
+substr(x, start, stop)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column.</p>
+</td></tr>
+<tr valign="top"><td><code>start</code></td>
+<td>
+<p>starting position.</p>
+</td></tr>
+<tr valign="top"><td><code>stop</code></td>
+<td>
+<p>ending position.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>substr since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="alias.html">alias</a></code>,
+<code><a href="between.html">between</a></code>, <code><a href="cast.html">cast</a></code>,
+<code><a href="endsWith.html">endsWith</a></code>, <code><a href="otherwise.html">otherwise</a></code>,
+<code><a href="over.html">over</a></code>, <code><a href="startsWith.html">startsWith</a></code>
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/26c57a24/site/docs/2.3.0/api/R/summarize.html
----------------------------------------------------------------------
diff --git a/site/docs/2.3.0/api/R/summarize.html b/site/docs/2.3.0/api/R/summarize.html
new file mode 100644
index 0000000..8f28639
--- /dev/null
+++ b/site/docs/2.3.0/api/R/summarize.html
@@ -0,0 +1,143 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: summarize</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for agg {SparkR}"><tr><td>agg {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>summarize</h2>
+
+<h3>Description</h3>
+
+<p>Aggregates on the entire SparkDataFrame without groups.
+The resulting SparkDataFrame will also contain the grouping columns.
+</p>
+<p>Compute aggregates by specifying a list of columns
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+agg(x, ...)
+
+summarize(x, ...)
+
+## S4 method for signature 'GroupedData'
+agg(x, ...)
+
+## S4 method for signature 'GroupedData'
+summarize(x, ...)
+
+## S4 method for signature 'SparkDataFrame'
+agg(x, ...)
+
+## S4 method for signature 'SparkDataFrame'
+summarize(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame or GroupedData.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>further arguments to be passed to or from other methods.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>df2 &lt;- agg(df, &lt;column&gt; = &lt;aggFunction&gt;)
+df2 &lt;- agg(df, newColName = aggFunction(column))
+</p>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>agg since 1.4.0
+</p>
+<p>summarize since 1.4.0
+</p>
+<p>agg since 1.4.0
+</p>
+<p>summarize since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="alias.html">alias</a></code>, <code><a href="arrange.html">arrange</a></code>,
+<code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="except.html">except</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="filter.html">filter</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapply.html">gapply</a></code>, <code><a href="getNumPartitions.html">getNumPartitions</a></code>,
+<code><a href="groupBy.html">group_by</a></code>, <code><a href="head.html">head</a></code>,
+<code><a href="hint.html">hint</a></code>, <code><a href="histogram.html">histogram</a></code>,
+<code><a href="insertInto.html">insertInto</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartition.html">repartition</a></code>,
+<code><a href="rollup.html">rollup</a></code>, <code><a href="sample.html">sample</a></code>,
+<code><a href="saveAsTable.html">saveAsTable</a></code>, <code><a href="schema.html">schema</a></code>,
+<code><a href="selectExpr.html">selectExpr</a></code>, <code><a href="select.html">select</a></code>,
+<code><a href="showDF.html">showDF</a></code>, <code><a href="show.html">show</a></code>,
+<code><a href="storageLevel.html">storageLevel</a></code>, <code><a href="str.html">str</a></code>,
+<code><a href="subset.html">subset</a></code>, <code><a href="summary.html">summary</a></code>,
+<code><a href="take.html">take</a></code>, <code><a href="toJSON.html">toJSON</a></code>,
+<code><a href="unionByName.html">unionByName</a></code>, <code><a href="union.html">union</a></code>,
+<code><a href="unpersist.html">unpersist</a></code>, <code><a href="withColumn.html">withColumn</a></code>,
+<code><a href="withWatermark.html">withWatermark</a></code>, <code><a href="with.html">with</a></code>,
+<code><a href="write.df.html">write.df</a></code>, <code><a href="write.jdbc.html">write.jdbc</a></code>,
+<code><a href="write.json.html">write.json</a></code>, <code><a href="write.orc.html">write.orc</a></code>,
+<code><a href="write.parquet.html">write.parquet</a></code>, <code><a href="write.stream.html">write.stream</a></code>,
+<code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D  df2 &lt;- agg(df, age = &quot;sum&quot;)  # new column name will be created as &#39;SUM(age#0)&#39;
+##D  df3 &lt;- agg(df, ageSum = sum(df$age)) # Creates a new column named ageSum
+##D  df4 &lt;- summarize(df, ageSum = max(df$age))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.3.0 <a href="00Index.html">Index</a>]</div>
+</body></html>


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org