You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2016/11/11 22:03:17 UTC

[28/51] [partial] spark-website git commit: Add docs for 2.0.2.

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/spark.lapply.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/spark.lapply.html b/site/docs/2.0.2/api/R/spark.lapply.html
new file mode 100644
index 0000000..f337327
--- /dev/null
+++ b/site/docs/2.0.2/api/R/spark.lapply.html
@@ -0,0 +1,96 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Run a function over a list of elements, distributing the...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for spark.lapply {SparkR}"><tr><td>spark.lapply {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Run a function over a list of elements, distributing the computations with Spark</h2>
+
+<h3>Description</h3>
+
+<p>Run a function over a list of elements, distributing the computations with Spark. Applies a
+function in a manner that is similar to doParallel or lapply to elements of a list.
+The computations are distributed using Spark. It is conceptually the same as the following code:
+lapply(list, func)
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+spark.lapply(list, func)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>list</code></td>
+<td>
+<p>the list of elements</p>
+</td></tr>
+<tr valign="top"><td><code>func</code></td>
+<td>
+<p>a function that takes one argument.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Known limitations:
+</p>
+
+<ul>
+<li><p> variable scoping and capture: compared to R's rich support for variable resolutions,
+the distributed nature of SparkR limits how variables are resolved at runtime. All the
+variables that are available through lexical scoping are embedded in the closure of the
+function and available as read-only variables within the function. The environment variables
+should be stored into temporary variables outside the function, and not directly accessed
+within the function.
+</p>
+</li>
+<li><p> loading external packages: In order to use a package, you need to load it inside the
+closure. For example, if you rely on the MASS module, here is how you would use it:
+</p>
+<PRE>
+    train &lt;- function(hyperparam) {
+      library(MASS)
+      lm.ridge("y ~ x+z", data, lambda=hyperparam)
+      model
+    }
+  </PRE>
+</li></ul>
+
+
+
+<h3>Value</h3>
+
+<p>a list of results (the exact type being determined by the function)
+</p>
+
+
+<h3>Note</h3>
+
+<p>spark.lapply since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D doubled &lt;- spark.lapply(1:10, function(x){2 * x})
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/spark.naiveBayes.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/spark.naiveBayes.html b/site/docs/2.0.2/api/R/spark.naiveBayes.html
new file mode 100644
index 0000000..b4d60c2
--- /dev/null
+++ b/site/docs/2.0.2/api/R/spark.naiveBayes.html
@@ -0,0 +1,143 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Naive Bayes Models</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for spark.naiveBayes {SparkR}"><tr><td>spark.naiveBayes {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Naive Bayes Models</h2>
+
+<h3>Description</h3>
+
+<p><code>spark.naiveBayes</code> fits a Bernoulli naive Bayes model against a SparkDataFrame.
+Users can call <code>summary</code> to print a summary of the fitted model, <code>predict</code> to make
+predictions on new data, and <code>write.ml</code>/<code>read.ml</code> to save/load fitted models.
+Only categorical data is supported.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+spark.naiveBayes(data, formula, ...)
+
+## S4 method for signature 'NaiveBayesModel'
+predict(object, newData)
+
+## S4 method for signature 'NaiveBayesModel'
+summary(object, ...)
+
+## S4 method for signature 'SparkDataFrame,formula'
+spark.naiveBayes(data, formula,
+  smoothing = 1, ...)
+
+## S4 method for signature 'NaiveBayesModel,character'
+write.ml(object, path,
+  overwrite = FALSE)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>data</code></td>
+<td>
+<p>a <code>SparkDataFrame</code> of observations and labels for model fitting.</p>
+</td></tr>
+<tr valign="top"><td><code>formula</code></td>
+<td>
+<p>a symbolic description of the model to be fitted. Currently only a few formula
+operators are supported, including '~', '.', ':', '+', and '-'.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s) passed to the method. Currently only <code>smoothing</code>.</p>
+</td></tr>
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a naive Bayes model fitted by <code>spark.naiveBayes</code>.</p>
+</td></tr>
+<tr valign="top"><td><code>newData</code></td>
+<td>
+<p>a SparkDataFrame for testing.</p>
+</td></tr>
+<tr valign="top"><td><code>smoothing</code></td>
+<td>
+<p>smoothing parameter.</p>
+</td></tr>
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>the directory where the model is saved</p>
+</td></tr>
+<tr valign="top"><td><code>overwrite</code></td>
+<td>
+<p>overwrites or not if the output path already exists. Default is FALSE
+which means throw exception if the output path exists.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p><code>predict</code> returns a SparkDataFrame containing predicted labeled in a column named
+&quot;prediction&quot;
+</p>
+<p><code>summary</code> returns a list containing <code>apriori</code>, the label distribution, and
+<code>tables</code>, conditional probabilities given the target label.
+</p>
+<p><code>spark.naiveBayes</code> returns a fitted naive Bayes model.
+</p>
+
+
+<h3>Note</h3>
+
+<p>predict(NaiveBayesModel) since 2.0.0
+</p>
+<p>summary(NaiveBayesModel) since 2.0.0
+</p>
+<p>spark.naiveBayes since 2.0.0
+</p>
+<p>write.ml(NaiveBayesModel, character) since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>e1071: <a href="https://cran.r-project.org/package=e1071">https://cran.r-project.org/package=e1071</a>
+</p>
+<p><a href="read.ml.html">read.ml</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D data &lt;- as.data.frame(UCBAdmissions)
+##D df &lt;- createDataFrame(data)
+##D 
+##D # fit a Bernoulli naive Bayes model
+##D model &lt;- spark.naiveBayes(df, Admit ~ Gender + Dept, smoothing = 0)
+##D 
+##D # get the summary of the model
+##D summary(model)
+##D 
+##D # make predictions
+##D predictions &lt;- predict(model, df)
+##D 
+##D # save and load the model
+##D path &lt;- &quot;path/to/model&quot;
+##D write.ml(model, path)
+##D savedModel &lt;- read.ml(path)
+##D summary(savedModel)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/spark.survreg.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/spark.survreg.html b/site/docs/2.0.2/api/R/spark.survreg.html
new file mode 100644
index 0000000..3bb08fc
--- /dev/null
+++ b/site/docs/2.0.2/api/R/spark.survreg.html
@@ -0,0 +1,133 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Accelerated Failure Time (AFT) Survival Regression Model</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for spark.survreg {SparkR}"><tr><td>spark.survreg {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Accelerated Failure Time (AFT) Survival Regression Model</h2>
+
+<h3>Description</h3>
+
+<p><code>spark.survreg</code> fits an accelerated failure time (AFT) survival regression model on
+a SparkDataFrame. Users can call <code>summary</code> to get a summary of the fitted AFT model,
+<code>predict</code> to make predictions on new data, and <code>write.ml</code>/<code>read.ml</code> to
+save/load fitted models.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+spark.survreg(data, formula)
+
+## S4 method for signature 'AFTSurvivalRegressionModel,character'
+write.ml(object, path,
+  overwrite = FALSE)
+
+## S4 method for signature 'SparkDataFrame,formula'
+spark.survreg(data, formula)
+
+## S4 method for signature 'AFTSurvivalRegressionModel'
+summary(object)
+
+## S4 method for signature 'AFTSurvivalRegressionModel'
+predict(object, newData)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>data</code></td>
+<td>
+<p>a SparkDataFrame for training.</p>
+</td></tr>
+<tr valign="top"><td><code>formula</code></td>
+<td>
+<p>a symbolic description of the model to be fitted. Currently only a few formula
+operators are supported, including '~', ':', '+', and '-'.
+Note that operator '.' is not supported currently.</p>
+</td></tr>
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a fitted AFT survival regression model.</p>
+</td></tr>
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>the directory where the model is saved.</p>
+</td></tr>
+<tr valign="top"><td><code>overwrite</code></td>
+<td>
+<p>overwrites or not if the output path already exists. Default is FALSE
+which means throw exception if the output path exists.</p>
+</td></tr>
+<tr valign="top"><td><code>newData</code></td>
+<td>
+<p>a SparkDataFrame for testing.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p><code>spark.survreg</code> returns a fitted AFT survival regression model.
+</p>
+<p><code>summary</code> returns a list containing the model's coefficients,
+intercept and log(scale)
+</p>
+<p><code>predict</code> returns a SparkDataFrame containing predicted values
+on the original scale of the data (mean predicted value at scale = 1.0).
+</p>
+
+
+<h3>Note</h3>
+
+<p>write.ml(AFTSurvivalRegressionModel, character) since 2.0.0
+</p>
+<p>spark.survreg since 2.0.0
+</p>
+<p>summary(AFTSurvivalRegressionModel) since 2.0.0
+</p>
+<p>predict(AFTSurvivalRegressionModel) since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="read.ml.html">read.ml</a>
+</p>
+<p>survival: <a href="https://cran.r-project.org/package=survival">https://cran.r-project.org/package=survival</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(ovarian)
+##D model &lt;- spark.survreg(df, Surv(futime, fustat) ~ ecog_ps + rx)
+##D 
+##D # get a summary of the model
+##D summary(model)
+##D 
+##D # make predictions
+##D predicted &lt;- predict(model, df)
+##D showDF(predicted)
+##D 
+##D # save and load the model
+##D path &lt;- &quot;path/to/model&quot;
+##D write.ml(model, path)
+##D savedModel &lt;- read.ml(path)
+##D summary(savedModel)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.callJMethod.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.callJMethod.html b/site/docs/2.0.2/api/R/sparkR.callJMethod.html
new file mode 100644
index 0000000..8f5706d
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.callJMethod.html
@@ -0,0 +1,91 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Call Java Methods</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.callJMethod {SparkR}"><tr><td>sparkR.callJMethod {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Call Java Methods</h2>
+
+<h3>Description</h3>
+
+<p>Call a Java method in the JVM running the Spark driver. The return
+values are automatically converted to R objects for simple objects. Other
+values are returned as &quot;jobj&quot; which are references to objects on JVM.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.callJMethod(x, methodName, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>object to invoke the method on. Should be a &quot;jobj&quot; created by newJObject.</p>
+</td></tr>
+<tr valign="top"><td><code>methodName</code></td>
+<td>
+<p>method name to call.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>parameters to pass to the Java method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is a low level function to access the JVM directly and should only be used
+for advanced use cases. The arguments and return values that are primitive R
+types (like integer, numeric, character, lists) are automatically translated to/from
+Java types (like Integer, Double, String, Array). A full list can be found in
+serialize.R and deserialize.R in the Apache Spark code base.
+</p>
+
+
+<h3>Value</h3>
+
+<p>the return value of the Java method. Either returned as a R object
+if it can be deserialized or returned as a &quot;jobj&quot;. See details section for more.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.callJMethod since 2.0.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.callJStatic.html">sparkR.callJStatic</a>, <a href="sparkR.newJObject.html">sparkR.newJObject</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session() # Need to have a Spark JVM running before calling newJObject
+##D # Create a Java ArrayList and populate it
+##D jarray &lt;- sparkR.newJObject(&quot;java.util.ArrayList&quot;)
+##D sparkR.callJMethod(jarray, &quot;add&quot;, 42L)
+##D sparkR.callJMethod(jarray, &quot;get&quot;, 0L) # Will print 42
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.callJStatic.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.callJStatic.html b/site/docs/2.0.2/api/R/sparkR.callJStatic.html
new file mode 100644
index 0000000..d567eb6
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.callJStatic.html
@@ -0,0 +1,89 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Call Static Java Methods</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.callJStatic {SparkR}"><tr><td>sparkR.callJStatic {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Call Static Java Methods</h2>
+
+<h3>Description</h3>
+
+<p>Call a static method in the JVM running the Spark driver. The return
+value is automatically converted to R objects for simple objects. Other
+values are returned as &quot;jobj&quot; which are references to objects on JVM.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.callJStatic(x, methodName, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>fully qualified Java class name that contains the static method to invoke.</p>
+</td></tr>
+<tr valign="top"><td><code>methodName</code></td>
+<td>
+<p>name of static method to invoke.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>parameters to pass to the Java method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is a low level function to access the JVM directly and should only be used
+for advanced use cases. The arguments and return values that are primitive R
+types (like integer, numeric, character, lists) are automatically translated to/from
+Java types (like Integer, Double, String, Array). A full list can be found in
+serialize.R and deserialize.R in the Apache Spark code base.
+</p>
+
+
+<h3>Value</h3>
+
+<p>the return value of the Java method. Either returned as a R object
+if it can be deserialized or returned as a &quot;jobj&quot;. See details section for more.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.callJStatic since 2.0.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.callJMethod.html">sparkR.callJMethod</a>, <a href="sparkR.newJObject.html">sparkR.newJObject</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session() # Need to have a Spark JVM running before calling callJStatic
+##D sparkR.callJStatic(&quot;java.lang.System&quot;, &quot;currentTimeMillis&quot;)
+##D sparkR.callJStatic(&quot;java.lang.System&quot;, &quot;getProperty&quot;, &quot;java.home&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.conf.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.conf.html b/site/docs/2.0.2/api/R/sparkR.conf.html
new file mode 100644
index 0000000..22db3ca
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.conf.html
@@ -0,0 +1,69 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Get Runtime Config from the current active SparkSession</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.conf {SparkR}"><tr><td>sparkR.conf {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Get Runtime Config from the current active SparkSession</h2>
+
+<h3>Description</h3>
+
+<p>Get Runtime Config from the current active SparkSession.
+To change SparkSession Runtime Config, please see <code>sparkR.session()</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.conf(key, defaultValue)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>key</code></td>
+<td>
+<p>(optional) The key of the config to get, if omitted, all config is returned</p>
+</td></tr>
+<tr valign="top"><td><code>defaultValue</code></td>
+<td>
+<p>(optional) The default value of the config to return if they config is not
+set, if omitted, the call fails if the config key is not set</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>a list of config values with keys as their names
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.conf since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D allConfigs &lt;- sparkR.conf()
+##D masterValue &lt;- unlist(sparkR.conf(&quot;spark.master&quot;))
+##D namedConfig &lt;- sparkR.conf(&quot;spark.executor.memory&quot;, &quot;0g&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.init-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.init-deprecated.html b/site/docs/2.0.2/api/R/sparkR.init-deprecated.html
new file mode 100644
index 0000000..0619f7a
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.init-deprecated.html
@@ -0,0 +1,93 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: (Deprecated) Initialize a new Spark Context</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.init {SparkR}"><tr><td>sparkR.init {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Initialize a new Spark Context</h2>
+
+<h3>Description</h3>
+
+<p>This function initializes a new SparkContext.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.init(master = "", appName = "SparkR",
+  sparkHome = Sys.getenv("SPARK_HOME"), sparkEnvir = list(),
+  sparkExecutorEnv = list(), sparkJars = "", sparkPackages = "")
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>master</code></td>
+<td>
+<p>The Spark master URL</p>
+</td></tr>
+<tr valign="top"><td><code>appName</code></td>
+<td>
+<p>Application name to register with cluster manager</p>
+</td></tr>
+<tr valign="top"><td><code>sparkHome</code></td>
+<td>
+<p>Spark Home directory</p>
+</td></tr>
+<tr valign="top"><td><code>sparkEnvir</code></td>
+<td>
+<p>Named list of environment variables to set on worker nodes</p>
+</td></tr>
+<tr valign="top"><td><code>sparkExecutorEnv</code></td>
+<td>
+<p>Named list of environment variables to be used when launching executors</p>
+</td></tr>
+<tr valign="top"><td><code>sparkJars</code></td>
+<td>
+<p>Character vector of jar files to pass to the worker nodes</p>
+</td></tr>
+<tr valign="top"><td><code>sparkPackages</code></td>
+<td>
+<p>Character vector of package coordinates</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>sparkR.init since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.session.html">sparkR.session</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sc &lt;- sparkR.init(&quot;local[2]&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;)
+##D sc &lt;- sparkR.init(&quot;local[2]&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;,
+##D                  list(spark.executor.memory=&quot;1g&quot;))
+##D sc &lt;- sparkR.init(&quot;yarn-client&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;,
+##D                  list(spark.executor.memory=&quot;4g&quot;),
+##D                  list(LD_LIBRARY_PATH=&quot;/directory of JVM libraries (libjvm.so) on workers/&quot;),
+##D                  c(&quot;one.jar&quot;, &quot;two.jar&quot;, &quot;three.jar&quot;),
+##D                  c(&quot;com.databricks:spark-avro_2.10:2.0.1&quot;))
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.newJObject.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.newJObject.html b/site/docs/2.0.2/api/R/sparkR.newJObject.html
new file mode 100644
index 0000000..9757f20
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.newJObject.html
@@ -0,0 +1,87 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Create Java Objects</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.newJObject {SparkR}"><tr><td>sparkR.newJObject {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Create Java Objects</h2>
+
+<h3>Description</h3>
+
+<p>Create a new Java object in the JVM running the Spark driver. The return
+value is automatically converted to an R object for simple objects. Other
+values are returned as a &quot;jobj&quot; which is a reference to an object on JVM.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.newJObject(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>fully qualified Java class name.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>arguments to be passed to the constructor.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is a low level function to access the JVM directly and should only be used
+for advanced use cases. The arguments and return values that are primitive R
+types (like integer, numeric, character, lists) are automatically translated to/from
+Java types (like Integer, Double, String, Array). A full list can be found in
+serialize.R and deserialize.R in the Apache Spark code base.
+</p>
+
+
+<h3>Value</h3>
+
+<p>the object created. Either returned as a R object
+if it can be deserialized or returned as a &quot;jobj&quot;. See details section for more.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.newJObject since 2.0.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.callJMethod.html">sparkR.callJMethod</a>, <a href="sparkR.callJStatic.html">sparkR.callJStatic</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session() # Need to have a Spark JVM running before calling newJObject
+##D # Create a Java ArrayList and populate it
+##D jarray &lt;- sparkR.newJObject(&quot;java.util.ArrayList&quot;)
+##D sparkR.callJMethod(jarray, &quot;add&quot;, 42L)
+##D sparkR.callJMethod(jarray, &quot;get&quot;, 0L) # Will print 42
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.session.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.session.html b/site/docs/2.0.2/api/R/sparkR.session.html
new file mode 100644
index 0000000..322bb28
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.session.html
@@ -0,0 +1,103 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Get the existing SparkSession or initialize a new...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.session {SparkR}"><tr><td>sparkR.session {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Get the existing SparkSession or initialize a new SparkSession.</h2>
+
+<h3>Description</h3>
+
+<p>SparkSession is the entry point into SparkR. <code>sparkR.session</code> gets the existing
+SparkSession or initializes a new SparkSession.
+Additional Spark properties can be set in <code>...</code>, and these named parameters take priority
+over values in <code>master</code>, <code>appName</code>, named lists of <code>sparkConfig</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.session(master = "", appName = "SparkR",
+  sparkHome = Sys.getenv("SPARK_HOME"), sparkConfig = list(),
+  sparkJars = "", sparkPackages = "", enableHiveSupport = TRUE, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>master</code></td>
+<td>
+<p>the Spark master URL.</p>
+</td></tr>
+<tr valign="top"><td><code>appName</code></td>
+<td>
+<p>application name to register with cluster manager.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkHome</code></td>
+<td>
+<p>Spark Home directory.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkConfig</code></td>
+<td>
+<p>named list of Spark configuration to set on worker nodes.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkJars</code></td>
+<td>
+<p>character vector of jar files to pass to the worker nodes.</p>
+</td></tr>
+<tr valign="top"><td><code>sparkPackages</code></td>
+<td>
+<p>character vector of package coordinates</p>
+</td></tr>
+<tr valign="top"><td><code>enableHiveSupport</code></td>
+<td>
+<p>enable support for Hive, fallback if not built with Hive support; once
+set, this cannot be turned off on an existing session</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>named Spark properties passed to the method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>For details on how to initialize and use SparkR, refer to SparkR programming guide at
+<a href="http://spark.apache.org/docs/latest/sparkr.html#starting-up-sparksession">http://spark.apache.org/docs/latest/sparkr.html#starting-up-sparksession</a>.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.session since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df &lt;- read.json(path)
+##D 
+##D sparkR.session(&quot;local[2]&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;)
+##D sparkR.session(&quot;yarn-client&quot;, &quot;SparkR&quot;, &quot;/home/spark&quot;,
+##D                list(spark.executor.memory=&quot;4g&quot;),
+##D                c(&quot;one.jar&quot;, &quot;two.jar&quot;, &quot;three.jar&quot;),
+##D                c(&quot;com.databricks:spark-avro_2.10:2.0.1&quot;))
+##D sparkR.session(spark.master = &quot;yarn-client&quot;, spark.executor.memory = &quot;4g&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.session.stop.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.session.stop.html b/site/docs/2.0.2/api/R/sparkR.session.stop.html
new file mode 100644
index 0000000..2d3ea35
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.session.stop.html
@@ -0,0 +1,40 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Stop the Spark Session and Spark Context</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+</head><body>
+
+<table width="100%" summary="page for sparkR.session.stop {SparkR}"><tr><td>sparkR.session.stop {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Stop the Spark Session and Spark Context</h2>
+
+<h3>Description</h3>
+
+<p>Stop the Spark Session and Spark Context.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.session.stop()
+
+sparkR.stop()
+</pre>
+
+
+<h3>Details</h3>
+
+<p>Also terminates the backend this R session is connected to.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.session.stop since 2.0.0
+</p>
+<p>sparkR.stop since 1.4.0
+</p>
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkR.version.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkR.version.html b/site/docs/2.0.2/api/R/sparkR.version.html
new file mode 100644
index 0000000..84006a9
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkR.version.html
@@ -0,0 +1,51 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Get version of Spark on which this application is running</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkR.version {SparkR}"><tr><td>sparkR.version {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Get version of Spark on which this application is running</h2>
+
+<h3>Description</h3>
+
+<p>Get version of Spark on which this application is running.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkR.version()
+</pre>
+
+
+<h3>Value</h3>
+
+<p>a character string of the Spark version
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkR.version since 2.0.1
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D version &lt;- sparkR.version()
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkRHive.init-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkRHive.init-deprecated.html b/site/docs/2.0.2/api/R/sparkRHive.init-deprecated.html
new file mode 100644
index 0000000..0bc7e5e
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkRHive.init-deprecated.html
@@ -0,0 +1,68 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: (Deprecated) Initialize a new HiveContext</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkRHive.init {SparkR}"><tr><td>sparkRHive.init {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Initialize a new HiveContext</h2>
+
+<h3>Description</h3>
+
+<p>This function creates a HiveContext from an existing JavaSparkContext
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkRHive.init(jsc = NULL)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jsc</code></td>
+<td>
+<p>The existing JavaSparkContext created with SparkR.init()</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Starting SparkR 2.0, a SparkSession is initialized and returned instead.
+This API is deprecated and kept for backward compatibility only.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkRHive.init since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.session.html">sparkR.session</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sc &lt;- sparkR.init()
+##D sqlContext &lt;- sparkRHive.init(sc)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sparkRSQL.init-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sparkRSQL.init-deprecated.html b/site/docs/2.0.2/api/R/sparkRSQL.init-deprecated.html
new file mode 100644
index 0000000..3cc4568
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sparkRSQL.init-deprecated.html
@@ -0,0 +1,69 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: (Deprecated) Initialize a new SQLContext</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sparkRSQL.init {SparkR}"><tr><td>sparkRSQL.init {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Initialize a new SQLContext</h2>
+
+<h3>Description</h3>
+
+<p>This function creates a SparkContext from an existing JavaSparkContext and
+then uses it to initialize a new SQLContext
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+sparkRSQL.init(jsc = NULL)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jsc</code></td>
+<td>
+<p>The existing JavaSparkContext created with SparkR.init()</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Starting SparkR 2.0, a SparkSession is initialized and returned instead.
+This API is deprecated and kept for backward compatibility only.
+</p>
+
+
+<h3>Note</h3>
+
+<p>sparkRSQL.init since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sparkR.session.html">sparkR.session</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sc &lt;- sparkR.init()
+##D sqlContext &lt;- sparkRSQL.init(sc)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/spark_partition_id.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/spark_partition_id.html b/site/docs/2.0.2/api/R/spark_partition_id.html
new file mode 100644
index 0000000..59ad9fe
--- /dev/null
+++ b/site/docs/2.0.2/api/R/spark_partition_id.html
@@ -0,0 +1,63 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Return the partition ID as a column</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for spark_partition_id {SparkR}"><tr><td>spark_partition_id {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Return the partition ID as a column</h2>
+
+<h3>Description</h3>
+
+<p>Return the partition ID of the Spark task as a SparkDataFrame column.
+Note that this is nondeterministic because it depends on data partitioning and
+task scheduling.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'missing'
+spark_partition_id()
+
+spark_partition_id(x = "missing")
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>empty. Should be used with no argument.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>This is equivalent to the SPARK_PARTITION_ID function in SQL.
+</p>
+
+
+<h3>Note</h3>
+
+<p>spark_partition_id since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: select(df, spark_partition_id())
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sql.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sql.html b/site/docs/2.0.2/api/R/sql.html
new file mode 100644
index 0000000..3f7d2ef
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sql.html
@@ -0,0 +1,65 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: SQL Query</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sql {SparkR}"><tr><td>sql {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>SQL Query</h2>
+
+<h3>Description</h3>
+
+<p>Executes a SQL query using Spark, returning the result as a SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+sql(sqlQuery)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>sqlQuery</code></td>
+<td>
+<p>A character vector containing the SQL query</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>sql since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D createOrReplaceTempView(df, &quot;table&quot;)
+##D new_df &lt;- sql(&quot;SELECT * FROM table&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/sqrt.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/sqrt.html b/site/docs/2.0.2/api/R/sqrt.html
new file mode 100644
index 0000000..1fec1eb
--- /dev/null
+++ b/site/docs/2.0.2/api/R/sqrt.html
@@ -0,0 +1,120 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: sqrt</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for sqrt {SparkR}"><tr><td>sqrt {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>sqrt</h2>
+
+<h3>Description</h3>
+
+<p>Computes the square root of the specified float value.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'Column'
+sqrt(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>sqrt since 1.5.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other math_funcs: <code><a href="acos.html">acos</a></code>,
+<code><a href="acos.html">acos,Column-method</a></code>; <code><a href="asin.html">asin</a></code>,
+<code><a href="asin.html">asin,Column-method</a></code>; <code><a href="atan2.html">atan2</a></code>,
+<code><a href="atan2.html">atan2,Column-method</a></code>; <code><a href="atan.html">atan</a></code>,
+<code><a href="atan.html">atan,Column-method</a></code>; <code><a href="bin.html">bin</a></code>,
+<code><a href="bin.html">bin</a></code>, <code><a href="bin.html">bin,Column-method</a></code>;
+<code><a href="bround.html">bround</a></code>, <code><a href="bround.html">bround</a></code>,
+<code><a href="bround.html">bround,Column-method</a></code>; <code><a href="cbrt.html">cbrt</a></code>,
+<code><a href="cbrt.html">cbrt</a></code>, <code><a href="cbrt.html">cbrt,Column-method</a></code>;
+<code><a href="ceil.html">ceil</a></code>, <code><a href="ceil.html">ceil</a></code>,
+<code><a href="ceil.html">ceil,Column-method</a></code>, <code><a href="ceil.html">ceiling</a></code>,
+<code><a href="ceil.html">ceiling,Column-method</a></code>; <code><a href="conv.html">conv</a></code>,
+<code><a href="conv.html">conv</a></code>,
+<code><a href="conv.html">conv,Column,numeric,numeric-method</a></code>;
+<code><a href="corr.html">corr</a></code>, <code><a href="corr.html">corr</a></code>,
+<code><a href="corr.html">corr</a></code>, <code><a href="corr.html">corr,Column-method</a></code>,
+<code><a href="corr.html">corr,SparkDataFrame-method</a></code>;
+<code><a href="cosh.html">cosh</a></code>, <code><a href="cosh.html">cosh,Column-method</a></code>;
+<code><a href="cos.html">cos</a></code>, <code><a href="cos.html">cos,Column-method</a></code>;
+<code><a href="covar_pop.html">covar_pop</a></code>, <code><a href="covar_pop.html">covar_pop</a></code>,
+<code><a href="covar_pop.html">covar_pop,characterOrColumn,characterOrColumn-method</a></code>;
+<code><a href="cov.html">cov</a></code>, <code><a href="cov.html">cov</a></code>, <code><a href="cov.html">cov</a></code>,
+<code><a href="cov.html">cov,SparkDataFrame-method</a></code>,
+<code><a href="cov.html">cov,characterOrColumn-method</a></code>,
+<code><a href="cov.html">covar_samp</a></code>, <code><a href="cov.html">covar_samp</a></code>,
+<code><a href="cov.html">covar_samp,characterOrColumn,characterOrColumn-method</a></code>;
+<code><a href="expm1.html">expm1</a></code>, <code><a href="expm1.html">expm1,Column-method</a></code>;
+<code><a href="exp.html">exp</a></code>, <code><a href="exp.html">exp,Column-method</a></code>;
+<code><a href="factorial.html">factorial</a></code>,
+<code><a href="factorial.html">factorial,Column-method</a></code>;
+<code><a href="floor.html">floor</a></code>, <code><a href="floor.html">floor,Column-method</a></code>;
+<code><a href="hex.html">hex</a></code>, <code><a href="hex.html">hex</a></code>,
+<code><a href="hex.html">hex,Column-method</a></code>; <code><a href="hypot.html">hypot</a></code>,
+<code><a href="hypot.html">hypot</a></code>, <code><a href="hypot.html">hypot,Column-method</a></code>;
+<code><a href="log10.html">log10</a></code>, <code><a href="log10.html">log10,Column-method</a></code>;
+<code><a href="log1p.html">log1p</a></code>, <code><a href="log1p.html">log1p,Column-method</a></code>;
+<code><a href="log2.html">log2</a></code>, <code><a href="log2.html">log2,Column-method</a></code>;
+<code><a href="log.html">log</a></code>, <code><a href="log.html">log,Column-method</a></code>;
+<code><a href="pmod.html">pmod</a></code>, <code><a href="pmod.html">pmod</a></code>,
+<code><a href="pmod.html">pmod,Column-method</a></code>; <code><a href="rint.html">rint</a></code>,
+<code><a href="rint.html">rint</a></code>, <code><a href="rint.html">rint,Column-method</a></code>;
+<code><a href="round.html">round</a></code>, <code><a href="round.html">round,Column-method</a></code>;
+<code><a href="shiftLeft.html">shiftLeft</a></code>, <code><a href="shiftLeft.html">shiftLeft</a></code>,
+<code><a href="shiftLeft.html">shiftLeft,Column,numeric-method</a></code>;
+<code><a href="shiftRightUnsigned.html">shiftRightUnsigned</a></code>,
+<code><a href="shiftRightUnsigned.html">shiftRightUnsigned</a></code>,
+<code><a href="shiftRightUnsigned.html">shiftRightUnsigned,Column,numeric-method</a></code>;
+<code><a href="shiftRight.html">shiftRight</a></code>, <code><a href="shiftRight.html">shiftRight</a></code>,
+<code><a href="shiftRight.html">shiftRight,Column,numeric-method</a></code>;
+<code><a href="sign.html">sign</a></code>, <code><a href="sign.html">sign,Column-method</a></code>,
+<code><a href="sign.html">signum</a></code>, <code><a href="sign.html">signum</a></code>,
+<code><a href="sign.html">signum,Column-method</a></code>; <code><a href="sinh.html">sinh</a></code>,
+<code><a href="sinh.html">sinh,Column-method</a></code>; <code><a href="sin.html">sin</a></code>,
+<code><a href="sin.html">sin,Column-method</a></code>; <code><a href="tanh.html">tanh</a></code>,
+<code><a href="tanh.html">tanh,Column-method</a></code>; <code><a href="tan.html">tan</a></code>,
+<code><a href="tan.html">tan,Column-method</a></code>; <code><a href="toDegrees.html">toDegrees</a></code>,
+<code><a href="toDegrees.html">toDegrees</a></code>,
+<code><a href="toDegrees.html">toDegrees,Column-method</a></code>;
+<code><a href="toRadians.html">toRadians</a></code>, <code><a href="toRadians.html">toRadians</a></code>,
+<code><a href="toRadians.html">toRadians,Column-method</a></code>;
+<code><a href="unhex.html">unhex</a></code>, <code><a href="unhex.html">unhex</a></code>,
+<code><a href="unhex.html">unhex,Column-method</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: sqrt(df$c)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/startsWith.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/startsWith.html b/site/docs/2.0.2/api/R/startsWith.html
new file mode 100644
index 0000000..e0cf767
--- /dev/null
+++ b/site/docs/2.0.2/api/R/startsWith.html
@@ -0,0 +1,65 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: startsWith</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+</head><body>
+
+<table width="100%" summary="page for startsWith {SparkR}"><tr><td>startsWith {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>startsWith</h2>
+
+<h3>Description</h3>
+
+<p>Determines if entries of x start with string (entries of) prefix respectively,
+where strings are recycled to common lengths.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'Column'
+startsWith(x, prefix)
+
+startsWith(x, prefix)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>vector of character string whose &quot;starts&quot; are considered</p>
+</td></tr>
+<tr valign="top"><td><code>prefix</code></td>
+<td>
+<p>character vector (often of length one)</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>startsWith since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="alias.html">alias</a></code>,
+<code><a href="alias.html">alias,Column-method</a></code>; <code><a href="between.html">between</a></code>,
+<code><a href="between.html">between</a></code>,
+<code><a href="between.html">between,Column-method</a></code>; <code><a href="cast.html">cast</a></code>,
+<code><a href="cast.html">cast</a></code>, <code><a href="cast.html">cast,Column-method</a></code>;
+<code><a href="endsWith.html">endsWith</a></code>, <code><a href="endsWith.html">endsWith</a></code>,
+<code><a href="endsWith.html">endsWith,Column-method</a></code>;
+<code><a href="otherwise.html">otherwise</a></code>, <code><a href="otherwise.html">otherwise</a></code>,
+<code><a href="otherwise.html">otherwise,Column-method</a></code>;
+<code><a href="over.html">over</a></code>, <code><a href="over.html">over</a></code>,
+<code><a href="over.html">over,Column,WindowSpec-method</a></code>;
+<code><a href="substr.html">substr</a></code>, <code><a href="substr.html">substr,Column-method</a></code>
+</p>
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/stddev_pop.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/stddev_pop.html b/site/docs/2.0.2/api/R/stddev_pop.html
new file mode 100644
index 0000000..b27882a
--- /dev/null
+++ b/site/docs/2.0.2/api/R/stddev_pop.html
@@ -0,0 +1,107 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: stddev_pop</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for stddev_pop {SparkR}"><tr><td>stddev_pop {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>stddev_pop</h2>
+
+<h3>Description</h3>
+
+<p>Aggregate function: returns the population standard deviation of the expression in a group.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'Column'
+stddev_pop(x)
+
+stddev_pop(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>stddev_pop since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="sd.html">sd</a>, <a href="stddev_samp.html">stddev_samp</a>
+</p>
+<p>Other agg_funcs: <code><a href="summarize.html">agg</a></code>, <code><a href="summarize.html">agg</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="summarize.html">agg,GroupedData-method</a></code>,
+<code><a href="summarize.html">agg,SparkDataFrame-method</a></code>,
+<code><a href="summarize.html">summarize</a></code>, <code><a href="summarize.html">summarize</a></code>,
+<code><a href="summarize.html">summarize</a></code>,
+<code><a href="summarize.html">summarize,GroupedData-method</a></code>,
+<code><a href="summarize.html">summarize,SparkDataFrame-method</a></code>;
+<code><a href="avg.html">avg</a></code>, <code><a href="avg.html">avg</a></code>,
+<code><a href="avg.html">avg,Column-method</a></code>;
+<code><a href="countDistinct.html">countDistinct</a></code>, <code><a href="countDistinct.html">countDistinct</a></code>,
+<code><a href="countDistinct.html">countDistinct,Column-method</a></code>,
+<code><a href="countDistinct.html">n_distinct</a></code>, <code><a href="countDistinct.html">n_distinct</a></code>,
+<code><a href="countDistinct.html">n_distinct,Column-method</a></code>;
+<code><a href="count.html">count</a></code>, <code><a href="count.html">count</a></code>,
+<code><a href="count.html">count,Column-method</a></code>,
+<code><a href="count.html">count,GroupedData-method</a></code>, <code><a href="count.html">n</a></code>,
+<code><a href="count.html">n</a></code>, <code><a href="count.html">n,Column-method</a></code>;
+<code><a href="first.html">first</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="first.html">first</a></code>,
+<code><a href="first.html">first,SparkDataFrame-method</a></code>,
+<code><a href="first.html">first,characterOrColumn-method</a></code>;
+<code><a href="kurtosis.html">kurtosis</a></code>, <code><a href="kurtosis.html">kurtosis</a></code>,
+<code><a href="kurtosis.html">kurtosis,Column-method</a></code>; <code><a href="last.html">last</a></code>,
+<code><a href="last.html">last</a></code>,
+<code><a href="last.html">last,characterOrColumn-method</a></code>;
+<code><a href="max.html">max</a></code>, <code><a href="max.html">max,Column-method</a></code>;
+<code><a href="mean.html">mean</a></code>, <code><a href="mean.html">mean,Column-method</a></code>;
+<code><a href="min.html">min</a></code>, <code><a href="min.html">min,Column-method</a></code>;
+<code><a href="sd.html">sd</a></code>, <code><a href="sd.html">sd</a></code>,
+<code><a href="sd.html">sd,Column-method</a></code>, <code><a href="sd.html">stddev</a></code>,
+<code><a href="sd.html">stddev</a></code>, <code><a href="sd.html">stddev,Column-method</a></code>;
+<code><a href="skewness.html">skewness</a></code>, <code><a href="skewness.html">skewness</a></code>,
+<code><a href="skewness.html">skewness,Column-method</a></code>;
+<code><a href="stddev_samp.html">stddev_samp</a></code>, <code><a href="stddev_samp.html">stddev_samp</a></code>,
+<code><a href="stddev_samp.html">stddev_samp,Column-method</a></code>;
+<code><a href="sumDistinct.html">sumDistinct</a></code>, <code><a href="sumDistinct.html">sumDistinct</a></code>,
+<code><a href="sumDistinct.html">sumDistinct,Column-method</a></code>;
+<code><a href="sum.html">sum</a></code>, <code><a href="sum.html">sum,Column-method</a></code>;
+<code><a href="var_pop.html">var_pop</a></code>, <code><a href="var_pop.html">var_pop</a></code>,
+<code><a href="var_pop.html">var_pop,Column-method</a></code>;
+<code><a href="var_samp.html">var_samp</a></code>, <code><a href="var_samp.html">var_samp</a></code>,
+<code><a href="var_samp.html">var_samp,Column-method</a></code>; <code><a href="var.html">var</a></code>,
+<code><a href="var.html">var</a></code>, <code><a href="var.html">var,Column-method</a></code>,
+<code><a href="var.html">variance</a></code>, <code><a href="var.html">variance</a></code>,
+<code><a href="var.html">variance,Column-method</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: stddev_pop(df$c)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/stddev_samp.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/stddev_samp.html b/site/docs/2.0.2/api/R/stddev_samp.html
new file mode 100644
index 0000000..8ed1ed6
--- /dev/null
+++ b/site/docs/2.0.2/api/R/stddev_samp.html
@@ -0,0 +1,107 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: stddev_samp</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for stddev_samp {SparkR}"><tr><td>stddev_samp {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>stddev_samp</h2>
+
+<h3>Description</h3>
+
+<p>Aggregate function: returns the unbiased sample standard deviation of the expression in a group.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'Column'
+stddev_samp(x)
+
+stddev_samp(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>stddev_samp since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="stddev_pop.html">stddev_pop</a>, <a href="sd.html">sd</a>
+</p>
+<p>Other agg_funcs: <code><a href="summarize.html">agg</a></code>, <code><a href="summarize.html">agg</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="summarize.html">agg,GroupedData-method</a></code>,
+<code><a href="summarize.html">agg,SparkDataFrame-method</a></code>,
+<code><a href="summarize.html">summarize</a></code>, <code><a href="summarize.html">summarize</a></code>,
+<code><a href="summarize.html">summarize</a></code>,
+<code><a href="summarize.html">summarize,GroupedData-method</a></code>,
+<code><a href="summarize.html">summarize,SparkDataFrame-method</a></code>;
+<code><a href="avg.html">avg</a></code>, <code><a href="avg.html">avg</a></code>,
+<code><a href="avg.html">avg,Column-method</a></code>;
+<code><a href="countDistinct.html">countDistinct</a></code>, <code><a href="countDistinct.html">countDistinct</a></code>,
+<code><a href="countDistinct.html">countDistinct,Column-method</a></code>,
+<code><a href="countDistinct.html">n_distinct</a></code>, <code><a href="countDistinct.html">n_distinct</a></code>,
+<code><a href="countDistinct.html">n_distinct,Column-method</a></code>;
+<code><a href="count.html">count</a></code>, <code><a href="count.html">count</a></code>,
+<code><a href="count.html">count,Column-method</a></code>,
+<code><a href="count.html">count,GroupedData-method</a></code>, <code><a href="count.html">n</a></code>,
+<code><a href="count.html">n</a></code>, <code><a href="count.html">n,Column-method</a></code>;
+<code><a href="first.html">first</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="first.html">first</a></code>,
+<code><a href="first.html">first,SparkDataFrame-method</a></code>,
+<code><a href="first.html">first,characterOrColumn-method</a></code>;
+<code><a href="kurtosis.html">kurtosis</a></code>, <code><a href="kurtosis.html">kurtosis</a></code>,
+<code><a href="kurtosis.html">kurtosis,Column-method</a></code>; <code><a href="last.html">last</a></code>,
+<code><a href="last.html">last</a></code>,
+<code><a href="last.html">last,characterOrColumn-method</a></code>;
+<code><a href="max.html">max</a></code>, <code><a href="max.html">max,Column-method</a></code>;
+<code><a href="mean.html">mean</a></code>, <code><a href="mean.html">mean,Column-method</a></code>;
+<code><a href="min.html">min</a></code>, <code><a href="min.html">min,Column-method</a></code>;
+<code><a href="sd.html">sd</a></code>, <code><a href="sd.html">sd</a></code>,
+<code><a href="sd.html">sd,Column-method</a></code>, <code><a href="sd.html">stddev</a></code>,
+<code><a href="sd.html">stddev</a></code>, <code><a href="sd.html">stddev,Column-method</a></code>;
+<code><a href="skewness.html">skewness</a></code>, <code><a href="skewness.html">skewness</a></code>,
+<code><a href="skewness.html">skewness,Column-method</a></code>;
+<code><a href="stddev_pop.html">stddev_pop</a></code>, <code><a href="stddev_pop.html">stddev_pop</a></code>,
+<code><a href="stddev_pop.html">stddev_pop,Column-method</a></code>;
+<code><a href="sumDistinct.html">sumDistinct</a></code>, <code><a href="sumDistinct.html">sumDistinct</a></code>,
+<code><a href="sumDistinct.html">sumDistinct,Column-method</a></code>;
+<code><a href="sum.html">sum</a></code>, <code><a href="sum.html">sum,Column-method</a></code>;
+<code><a href="var_pop.html">var_pop</a></code>, <code><a href="var_pop.html">var_pop</a></code>,
+<code><a href="var_pop.html">var_pop,Column-method</a></code>;
+<code><a href="var_samp.html">var_samp</a></code>, <code><a href="var_samp.html">var_samp</a></code>,
+<code><a href="var_samp.html">var_samp,Column-method</a></code>; <code><a href="var.html">var</a></code>,
+<code><a href="var.html">var</a></code>, <code><a href="var.html">var,Column-method</a></code>,
+<code><a href="var.html">variance</a></code>, <code><a href="var.html">variance</a></code>,
+<code><a href="var.html">variance,Column-method</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: stddev_samp(df$c)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/str.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/str.html b/site/docs/2.0.2/api/R/str.html
new file mode 100644
index 0000000..0989d9a
--- /dev/null
+++ b/site/docs/2.0.2/api/R/str.html
@@ -0,0 +1,259 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: Compactly display the structure of a dataset</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for str {SparkR}"><tr><td>str {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>Compactly display the structure of a dataset</h2>
+
+<h3>Description</h3>
+
+<p>Display the structure of a SparkDataFrame, including column names, column types, as well as a
+a small sample of rows.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'SparkDataFrame'
+str(object)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a SparkDataFrame</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>str since 1.6.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="select.html">$</a></code>,
+<code><a href="select.html">$,SparkDataFrame-method</a></code>, <code><a href="select.html">$&lt;-</a></code>,
+<code><a href="select.html">$&lt;-,SparkDataFrame-method</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="select.html">select</a></code>,
+<code><a href="select.html">select,SparkDataFrame,Column-method</a></code>,
+<code><a href="select.html">select,SparkDataFrame,character-method</a></code>,
+<code><a href="select.html">select,SparkDataFrame,list-method</a></code>;
+<code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>; <code><a href="subset.html">[</a></code>,
+<code><a href="subset.html">[,SparkDataFrame-method</a></code>, <code><a href="subset.html">[[</a></code>,
+<code><a href="subset.html">[[,SparkDataFrame,numericOrcharacter-method</a></code>,
+<code><a href="subset.html">subset</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="subset.html">subset,SparkDataFrame-method</a></code>;
+<code><a href="summarize.html">agg</a></code>, <code><a href="summarize.html">agg</a></code>, <code><a href="summarize.html">agg</a></code>,
+<code><a href="summarize.html">agg,GroupedData-method</a></code>,
+<code><a href="summarize.html">agg,SparkDataFrame-method</a></code>,
+<code><a href="summarize.html">summarize</a></code>, <code><a href="summarize.html">summarize</a></code>,
+<code><a href="summarize.html">summarize</a></code>,
+<code><a href="summarize.html">summarize,GroupedData-method</a></code>,
+<code><a href="summarize.html">summarize,SparkDataFrame-method</a></code>;
+<code><a href="arrange.html">arrange</a></code>, <code><a href="arrange.html">arrange</a></code>,
+<code><a href="arrange.html">arrange</a></code>,
+<code><a href="arrange.html">arrange,SparkDataFrame,Column-method</a></code>,
+<code><a href="arrange.html">arrange,SparkDataFrame,character-method</a></code>,
+<code><a href="arrange.html">orderBy,SparkDataFrame,characterOrColumn-method</a></code>;
+<code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="as.data.frame.html">as.data.frame,SparkDataFrame-method</a></code>;
+<code><a href="attach.html">attach</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>;
+<code><a href="cache.html">cache</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="cache.html">cache,SparkDataFrame-method</a></code>;
+<code><a href="collect.html">collect</a></code>, <code><a href="collect.html">collect</a></code>,
+<code><a href="collect.html">collect,SparkDataFrame-method</a></code>;
+<code><a href="columns.html">colnames</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="columns.html">colnames,SparkDataFrame-method</a></code>,
+<code><a href="columns.html">colnames&lt;-</a></code>, <code><a href="columns.html">colnames&lt;-</a></code>,
+<code><a href="columns.html">colnames&lt;-,SparkDataFrame-method</a></code>,
+<code><a href="columns.html">columns</a></code>, <code><a href="columns.html">columns</a></code>,
+<code><a href="columns.html">columns,SparkDataFrame-method</a></code>,
+<code><a href="columns.html">names</a></code>,
+<code><a href="columns.html">names,SparkDataFrame-method</a></code>,
+<code><a href="columns.html">names&lt;-</a></code>,
+<code><a href="columns.html">names&lt;-,SparkDataFrame-method</a></code>;
+<code><a href="coltypes.html">coltypes</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="coltypes.html">coltypes,SparkDataFrame-method</a></code>,
+<code><a href="coltypes.html">coltypes&lt;-</a></code>, <code><a href="coltypes.html">coltypes&lt;-</a></code>,
+<code><a href="coltypes.html">coltypes&lt;-,SparkDataFrame,character-method</a></code>;
+<code><a href="nrow.html">count,SparkDataFrame-method</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="nrow.html">nrow</a></code>,
+<code><a href="nrow.html">nrow,SparkDataFrame-method</a></code>;
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView,SparkDataFrame,character-method</a></code>;
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapplyCollect.html">dapplyCollect</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect,SparkDataFrame,function-method</a></code>;
+<code><a href="dapply.html">dapply</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="dapply.html">dapply,SparkDataFrame,function,structType-method</a></code>;
+<code><a href="summary.html">describe</a></code>, <code><a href="summary.html">describe</a></code>,
+<code><a href="summary.html">describe</a></code>,
+<code><a href="summary.html">describe,SparkDataFrame,ANY-method</a></code>,
+<code><a href="summary.html">describe,SparkDataFrame,character-method</a></code>,
+<code><a href="summary.html">describe,SparkDataFrame-method</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="summary.html">summary</a></code>,
+<code><a href="summary.html">summary,SparkDataFrame-method</a></code>;
+<code><a href="dim.html">dim</a></code>,
+<code><a href="dim.html">dim,SparkDataFrame-method</a></code>;
+<code><a href="distinct.html">distinct</a></code>, <code><a href="distinct.html">distinct</a></code>,
+<code><a href="distinct.html">distinct,SparkDataFrame-method</a></code>,
+<code><a href="distinct.html">unique</a></code>,
+<code><a href="distinct.html">unique,SparkDataFrame-method</a></code>;
+<code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="dropDuplicates.html">dropDuplicates,SparkDataFrame-method</a></code>;
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="nafunctions.html">dropna</a></code>,
+<code><a href="nafunctions.html">dropna,SparkDataFrame-method</a></code>,
+<code><a href="nafunctions.html">fillna</a></code>, <code><a href="nafunctions.html">fillna</a></code>,
+<code><a href="nafunctions.html">fillna,SparkDataFrame-method</a></code>,
+<code><a href="nafunctions.html">na.omit</a></code>, <code><a href="nafunctions.html">na.omit</a></code>,
+<code><a href="nafunctions.html">na.omit,SparkDataFrame-method</a></code>;
+<code><a href="drop.html">drop</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="drop.html">drop</a></code>, <code><a href="drop.html">drop,ANY-method</a></code>,
+<code><a href="drop.html">drop,SparkDataFrame-method</a></code>;
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="dtypes.html">dtypes</a></code>,
+<code><a href="dtypes.html">dtypes,SparkDataFrame-method</a></code>;
+<code><a href="except.html">except</a></code>, <code><a href="except.html">except</a></code>,
+<code><a href="except.html">except,SparkDataFrame,SparkDataFrame-method</a></code>;
+<code><a href="explain.html">explain</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="explain.html">explain,SparkDataFrame-method</a></code>;
+<code><a href="filter.html">filter</a></code>, <code><a href="filter.html">filter</a></code>,
+<code><a href="filter.html">filter,SparkDataFrame,characterOrColumn-method</a></code>,
+<code><a href="filter.html">where</a></code>, <code><a href="filter.html">where</a></code>,
+<code><a href="filter.html">where,SparkDataFrame,characterOrColumn-method</a></code>;
+<code><a href="first.html">first</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="first.html">first</a></code>,
+<code><a href="first.html">first,SparkDataFrame-method</a></code>,
+<code><a href="first.html">first,characterOrColumn-method</a></code>;
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect,GroupedData-method</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect,SparkDataFrame-method</a></code>;
+<code><a href="gapply.html">gapply</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="gapply.html">gapply</a></code>,
+<code><a href="gapply.html">gapply,GroupedData-method</a></code>,
+<code><a href="gapply.html">gapply,SparkDataFrame-method</a></code>;
+<code><a href="groupBy.html">groupBy</a></code>, <code><a href="groupBy.html">groupBy</a></code>,
+<code><a href="groupBy.html">groupBy,SparkDataFrame-method</a></code>,
+<code><a href="groupBy.html">group_by</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="groupBy.html">group_by,SparkDataFrame-method</a></code>;
+<code><a href="head.html">head</a></code>,
+<code><a href="head.html">head,SparkDataFrame-method</a></code>;
+<code><a href="histogram.html">histogram</a></code>,
+<code><a href="histogram.html">histogram,SparkDataFrame,characterOrColumn-method</a></code>;
+<code><a href="insertInto.html">insertInto</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="insertInto.html">insertInto,SparkDataFrame,character-method</a></code>;
+<code><a href="intersect.html">intersect</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="intersect.html">intersect,SparkDataFrame,SparkDataFrame-method</a></code>;
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isLocal.html">isLocal</a></code>,
+<code><a href="isLocal.html">isLocal,SparkDataFrame-method</a></code>;
+<code><a href="join.html">join</a></code>,
+<code><a href="join.html">join,SparkDataFrame,SparkDataFrame-method</a></code>;
+<code><a href="limit.html">limit</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="limit.html">limit,SparkDataFrame,numeric-method</a></code>;
+<code><a href="merge.html">merge</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="merge.html">merge,SparkDataFrame,SparkDataFrame-method</a></code>;
+<code><a href="mutate.html">mutate</a></code>, <code><a href="mutate.html">mutate</a></code>,
+<code><a href="mutate.html">mutate,SparkDataFrame-method</a></code>,
+<code><a href="mutate.html">transform</a></code>, <code><a href="mutate.html">transform</a></code>,
+<code><a href="mutate.html">transform,SparkDataFrame-method</a></code>;
+<code><a href="ncol.html">ncol</a></code>,
+<code><a href="ncol.html">ncol,SparkDataFrame-method</a></code>;
+<code><a href="persist.html">persist</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="persist.html">persist,SparkDataFrame,character-method</a></code>;
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="printSchema.html">printSchema</a></code>,
+<code><a href="printSchema.html">printSchema,SparkDataFrame-method</a></code>;
+<code><a href="randomSplit.html">randomSplit</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="randomSplit.html">randomSplit,SparkDataFrame,numeric-method</a></code>;
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rbind.html">rbind</a></code>,
+<code><a href="rbind.html">rbind,SparkDataFrame-method</a></code>;
+<code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
+<code><a href="registerTempTable-deprecated.html">registerTempTable</a></code>,
+<code><a href="registerTempTable-deprecated.html">registerTempTable,SparkDataFrame,character-method</a></code>;
+<code><a href="rename.html">rename</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="rename.html">rename,SparkDataFrame-method</a></code>,
+<code><a href="rename.html">withColumnRenamed</a></code>,
+<code><a href="rename.html">withColumnRenamed</a></code>,
+<code><a href="rename.html">withColumnRenamed,SparkDataFrame,character,character-method</a></code>;
+<code><a href="repartition.html">repartition</a></code>, <code><a href="repartition.html">repartition</a></code>,
+<code><a href="repartition.html">repartition,SparkDataFrame-method</a></code>;
+<code><a href="sample.html">sample</a></code>, <code><a href="sample.html">sample</a></code>,
+<code><a href="sample.html">sample,SparkDataFrame,logical,numeric-method</a></code>,
+<code><a href="sample.html">sample_frac</a></code>, <code><a href="sample.html">sample_frac</a></code>,
+<code><a href="sample.html">sample_frac,SparkDataFrame,logical,numeric-method</a></code>;
+<code><a href="write.parquet.html">saveAsParquetFile</a></code>,
+<code><a href="write.parquet.html">saveAsParquetFile</a></code>,
+<code><a href="write.parquet.html">saveAsParquetFile,SparkDataFrame,character-method</a></code>,
+<code><a href="write.parquet.html">write.parquet</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.parquet.html">write.parquet,SparkDataFrame,character-method</a></code>;
+<code><a href="saveAsTable.html">saveAsTable</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="saveAsTable.html">saveAsTable,SparkDataFrame,character-method</a></code>;
+<code><a href="write.df.html">saveDF</a></code>, <code><a href="write.df.html">saveDF</a></code>,
+<code><a href="write.df.html">saveDF,SparkDataFrame,character-method</a></code>,
+<code><a href="write.df.html">write.df</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.df.html">write.df,SparkDataFrame,character-method</a></code>;
+<code><a href="schema.html">schema</a></code>, <code><a href="schema.html">schema</a></code>,
+<code><a href="schema.html">schema,SparkDataFrame-method</a></code>;
+<code><a href="selectExpr.html">selectExpr</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="selectExpr.html">selectExpr,SparkDataFrame,character-method</a></code>;
+<code><a href="showDF.html">showDF</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="showDF.html">showDF,SparkDataFrame-method</a></code>;
+<code><a href="show.html">show</a></code>, <code><a href="show.html">show</a></code>,
+<code><a href="show.html">show,Column-method</a></code>,
+<code><a href="show.html">show,GroupedData-method</a></code>,
+<code><a href="show.html">show,SparkDataFrame-method</a></code>,
+<code><a href="show.html">show,WindowSpec-method</a></code>; <code><a href="take.html">take</a></code>,
+<code><a href="take.html">take</a></code>,
+<code><a href="take.html">take,SparkDataFrame,numeric-method</a></code>;
+<code><a href="union.html">union</a></code>, <code><a href="union.html">union</a></code>,
+<code><a href="union.html">union,SparkDataFrame,SparkDataFrame-method</a></code>,
+<code><a href="union.html">unionAll</a></code>, <code><a href="union.html">unionAll</a></code>,
+<code><a href="union.html">unionAll,SparkDataFrame,SparkDataFrame-method</a></code>;
+<code><a href="unpersist-methods.html">unpersist</a></code>, <code><a href="unpersist-methods.html">unpersist</a></code>,
+<code><a href="unpersist-methods.html">unpersist,SparkDataFrame-method</a></code>;
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withColumn.html">withColumn</a></code>,
+<code><a href="withColumn.html">withColumn,SparkDataFrame,character,Column-method</a></code>;
+<code><a href="with.html">with</a></code>,
+<code><a href="with.html">with,SparkDataFrame-method</a></code>;
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.jdbc.html">write.jdbc</a></code>,
+<code><a href="write.jdbc.html">write.jdbc,SparkDataFrame,character,character-method</a></code>;
+<code><a href="write.json.html">write.json</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.json.html">write.json,SparkDataFrame,character-method</a></code>;
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.orc.html">write.orc</a></code>,
+<code><a href="write.orc.html">write.orc,SparkDataFrame,character-method</a></code>;
+<code><a href="write.text.html">write.text</a></code>, <code><a href="write.text.html">write.text</a></code>,
+<code><a href="write.text.html">write.text,SparkDataFrame,character-method</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Create a SparkDataFrame from the Iris dataset
+##D irisDF &lt;- createDataFrame(iris)
+##D 
+##D # Show the structure of the SparkDataFrame
+##D str(irisDF)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0bd36316/site/docs/2.0.2/api/R/struct.html
----------------------------------------------------------------------
diff --git a/site/docs/2.0.2/api/R/struct.html b/site/docs/2.0.2/api/R/struct.html
new file mode 100644
index 0000000..2812eec
--- /dev/null
+++ b/site/docs/2.0.2/api/R/struct.html
@@ -0,0 +1,96 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
+<html><head><title>R: struct</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
+<link rel="stylesheet" type="text/css" href="R.css">
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for struct {SparkR}"><tr><td>struct {SparkR}</td><td align="right">R Documentation</td></tr></table>
+
+<h2>struct</h2>
+
+<h3>Description</h3>
+
+<p>Creates a new struct column that composes multiple input columns.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'characterOrColumn'
+struct(x, ...)
+
+struct(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>optional column(s) to be included.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>struct since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other normal_funcs: <code><a href="column.html">Column-class</a></code>,
+<code><a href="column.html">column</a></code>, <code><a href="column.html">column</a></code>,
+<code><a href="column.html">column</a></code>,
+<code><a href="column.html">column,character-method</a></code>,
+<code><a href="column.html">column,jobj-method</a></code>; <code><a href="abs.html">abs</a></code>,
+<code><a href="abs.html">abs,Column-method</a></code>;
+<code><a href="bitwiseNOT.html">bitwiseNOT</a></code>, <code><a href="bitwiseNOT.html">bitwiseNOT</a></code>,
+<code><a href="bitwiseNOT.html">bitwiseNOT,Column-method</a></code>;
+<code><a href="expr.html">expr</a></code>, <code><a href="expr.html">expr</a></code>,
+<code><a href="expr.html">expr,character-method</a></code>;
+<code><a href="greatest.html">greatest</a></code>, <code><a href="greatest.html">greatest</a></code>,
+<code><a href="greatest.html">greatest,Column-method</a></code>;
+<code><a href="ifelse.html">ifelse</a></code>, <code><a href="ifelse.html">ifelse,Column-method</a></code>;
+<code><a href="is.nan.html">is.nan</a></code>, <code><a href="is.nan.html">is.nan,Column-method</a></code>,
+<code><a href="is.nan.html">isnan</a></code>, <code><a href="is.nan.html">isnan</a></code>,
+<code><a href="is.nan.html">isnan,Column-method</a></code>; <code><a href="least.html">least</a></code>,
+<code><a href="least.html">least</a></code>, <code><a href="least.html">least,Column-method</a></code>;
+<code><a href="lit.html">lit</a></code>, <code><a href="lit.html">lit</a></code>,
+<code><a href="lit.html">lit,ANY-method</a></code>; <code><a href="nanvl.html">nanvl</a></code>,
+<code><a href="nanvl.html">nanvl</a></code>, <code><a href="nanvl.html">nanvl,Column-method</a></code>;
+<code><a href="negate.html">negate</a></code>, <code><a href="negate.html">negate</a></code>,
+<code><a href="negate.html">negate,Column-method</a></code>; <code><a href="randn.html">randn</a></code>,
+<code><a href="randn.html">randn</a></code>, <code><a href="randn.html">randn</a></code>,
+<code><a href="randn.html">randn,missing-method</a></code>,
+<code><a href="randn.html">randn,numeric-method</a></code>; <code><a href="rand.html">rand</a></code>,
+<code><a href="rand.html">rand</a></code>, <code><a href="rand.html">rand</a></code>,
+<code><a href="rand.html">rand,missing-method</a></code>,
+<code><a href="rand.html">rand,numeric-method</a></code>; <code><a href="when.html">when</a></code>,
+<code><a href="when.html">when</a></code>, <code><a href="when.html">when,Column-method</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D struct(df$c, df$d)
+##D struct(&quot;col1&quot;, &quot;col2&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr><div align="center">[Package <em>SparkR</em> version 2.0.2 <a href="00Index.html">Index</a>]</div>
+</body></html>


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org