You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by li...@apache.org on 2019/05/09 00:17:56 UTC

[spark-website] branch asf-site updated: Update Spark website for 2.4.3 release

This is an automated email from the ASF dual-hosted git repository.

lixiao pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/spark-website.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 0592e40  Update Spark website for 2.4.3 release
0592e40 is described below

commit 0592e40b1f6ea193fb97925874b5842c248bc537
Author: Xiao Li <ga...@gmail.com>
AuthorDate: Wed May 8 17:17:45 2019 -0700

    Update Spark website for 2.4.3 release
    
    This PR is to add the release note and news for Spark 2.4.3 release.
    
    Author: Xiao Li <ga...@gmail.com>
    
    Closes #201 from gatorsmile/updateWebsitefor2.4.3.
---
 _layouts/global.html                               |    2 +-
 documentation.md                                   |    1 +
 downloads.md                                       |    4 +-
 js/downloads.js                                    |    8 +-
 news/_posts/2019-05-08-spark-2-4-3-released.md     |   14 +
 releases/_posts/2019-05-08-spark-release-2-4-3.md  |   24 +
 site/committers.html                               |    8 +-
 site/community.html                                |    8 +-
 site/contributing.html                             |    8 +-
 site/developer-tools.html                          |    8 +-
 site/docs/2.4.3/api/R/00Index.html                 | 1928 +++++++++++++++++++-
 site/docs/2.4.3/api/R/00frame_toc.html             |  233 +++
 .../api/R/AFTSurvivalRegressionModel-class.html    |   32 +
 site/docs/2.4.3/api/R/ALSModel-class.html          |   32 +
 .../2.4.3/api/R/BisectingKMeansModel-class.html    |   32 +
 .../R/DecisionTreeClassificationModel-class.html   |   32 +
 .../api/R/DecisionTreeRegressionModel-class.html   |   32 +
 site/docs/2.4.3/api/R/FPGrowthModel-class.html     |   32 +
 .../2.4.3/api/R/GBTClassificationModel-class.html  |   32 +
 .../docs/2.4.3/api/R/GBTRegressionModel-class.html |   32 +
 .../2.4.3/api/R/GaussianMixtureModel-class.html    |   32 +
 .../R/GeneralizedLinearRegressionModel-class.html  |   32 +
 site/docs/2.4.3/api/R/GroupedData.html             |   45 +
 .../2.4.3/api/R/IsotonicRegressionModel-class.html |   32 +
 site/docs/2.4.3/api/R/KMeansModel-class.html       |   32 +
 site/docs/2.4.3/api/R/KSTest-class.html            |   32 +
 site/docs/2.4.3/api/R/LDAModel-class.html          |   32 +
 site/docs/2.4.3/api/R/LinearSVCModel-class.html    |   32 +
 .../2.4.3/api/R/LogisticRegressionModel-class.html |   32 +
 ...tilayerPerceptronClassificationModel-class.html |   32 +
 site/docs/2.4.3/api/R/NaiveBayesModel-class.html   |   32 +
 .../R/RandomForestClassificationModel-class.html   |   32 +
 .../api/R/RandomForestRegressionModel-class.html   |   32 +
 site/docs/2.4.3/api/R/SparkDataFrame.html          |  102 ++
 site/docs/2.4.3/api/R/StreamingQuery.html          |   40 +
 site/docs/2.4.3/api/R/WindowSpec.html              |   38 +
 site/docs/2.4.3/api/R/alias.html                   |  134 ++
 site/docs/2.4.3/api/R/approxQuantile.html          |   98 +
 site/docs/2.4.3/api/R/arrange.html                 |  147 ++
 site/docs/2.4.3/api/R/as.data.frame.html           |  127 ++
 site/docs/2.4.3/api/R/attach.html                  |  123 ++
 site/docs/2.4.3/api/R/avg.html                     |   67 +
 site/docs/2.4.3/api/R/awaitTermination.html        |   84 +
 site/docs/2.4.3/api/R/between.html                 |   55 +
 site/docs/2.4.3/api/R/broadcast.html               |  119 ++
 site/docs/2.4.3/api/R/cache.html                   |  107 ++
 site/docs/2.4.3/api/R/cacheTable.html              |   65 +
 site/docs/2.4.3/api/R/cancelJobGroup.html          |   55 +
 site/docs/2.4.3/api/R/cast.html                    |   72 +
 site/docs/2.4.3/api/R/checkpoint.html              |  120 ++
 site/docs/2.4.3/api/R/clearCache.html              |   44 +
 site/docs/2.4.3/api/R/clearJobGroup.html           |   45 +
 site/docs/2.4.3/api/R/coalesce.html                |  132 ++
 site/docs/2.4.3/api/R/collect.html                 |  118 ++
 site/docs/2.4.3/api/R/coltypes.html                |  133 ++
 site/docs/2.4.3/api/R/column.html                  |   77 +
 .../2.4.3/api/R/column_aggregate_functions.html    |  352 ++++
 .../2.4.3/api/R/column_collection_functions.html   |  503 +++++
 .../api/R/column_datetime_diff_functions.html      |  217 +++
 .../2.4.3/api/R/column_datetime_functions.html     |  403 ++++
 site/docs/2.4.3/api/R/column_math_functions.html   |  415 +++++
 site/docs/2.4.3/api/R/column_misc_functions.html   |  117 ++
 .../2.4.3/api/R/column_nonaggregate_functions.html |  351 ++++
 site/docs/2.4.3/api/R/column_string_functions.html |  522 ++++++
 site/docs/2.4.3/api/R/column_window_functions.html |  187 ++
 site/docs/2.4.3/api/R/columnfunctions.html         |   55 +
 site/docs/2.4.3/api/R/columns.html                 |  145 ++
 site/docs/2.4.3/api/R/corr.html                    |  110 ++
 site/docs/2.4.3/api/R/count.html                   |   89 +
 site/docs/2.4.3/api/R/cov.html                     |  137 ++
 site/docs/2.4.3/api/R/createDataFrame.html         |   90 +
 .../api/R/createExternalTable-deprecated.html      |   93 +
 site/docs/2.4.3/api/R/createOrReplaceTempView.html |  112 ++
 site/docs/2.4.3/api/R/createTable.html             |   91 +
 site/docs/2.4.3/api/R/crossJoin.html               |  117 ++
 site/docs/2.4.3/api/R/crosstab.html                |   82 +
 site/docs/2.4.3/api/R/cube.html                    |  128 ++
 site/docs/2.4.3/api/R/currentDatabase.html         |   50 +
 site/docs/2.4.3/api/R/dapply.html                  |  152 ++
 site/docs/2.4.3/api/R/dapplyCollect.html           |  130 ++
 site/docs/2.4.3/api/R/describe.html                |  131 ++
 site/docs/2.4.3/api/R/dim.html                     |  105 ++
 site/docs/2.4.3/api/R/distinct.html                |  112 ++
 site/docs/2.4.3/api/R/drop.html                    |  127 ++
 site/docs/2.4.3/api/R/dropDuplicates.html          |  121 ++
 .../docs/2.4.3/api/R/dropTempTable-deprecated.html |   64 +
 site/docs/2.4.3/api/R/dropTempView.html            |   63 +
 site/docs/2.4.3/api/R/dtypes.html                  |  107 ++
 site/docs/2.4.3/api/R/endsWith.html                |   56 +
 site/docs/2.4.3/api/R/eq_null_safe.html            |   76 +
 site/docs/2.4.3/api/R/except.html                  |  118 ++
 site/docs/2.4.3/api/R/exceptAll.html               |  120 ++
 site/docs/2.4.3/api/R/explain.html                 |  126 ++
 site/docs/2.4.3/api/R/filter.html                  |  129 ++
 site/docs/2.4.3/api/R/first.html                   |  139 ++
 site/docs/2.4.3/api/R/fitted.html                  |   74 +
 site/docs/2.4.3/api/R/freqItems.html               |   80 +
 site/docs/2.4.3/api/R/gapply.html                  |  208 +++
 site/docs/2.4.3/api/R/gapplyCollect.html           |  186 ++
 site/docs/2.4.3/api/R/getLocalProperty.html        |   55 +
 site/docs/2.4.3/api/R/getNumPartitions.html        |  104 ++
 site/docs/2.4.3/api/R/glm.html                     |  125 ++
 site/docs/2.4.3/api/R/groupBy.html                 |  127 ++
 site/docs/2.4.3/api/R/hashCode.html                |   68 +
 site/docs/2.4.3/api/R/head.html                    |  116 ++
 site/docs/2.4.3/api/R/hint.html                    |  121 ++
 site/docs/2.4.3/api/R/histogram.html               |  130 ++
 site/docs/2.4.3/api/R/insertInto.html              |  122 ++
 site/docs/2.4.3/api/R/install.spark.html           |  118 ++
 site/docs/2.4.3/api/R/intersect.html               |  118 ++
 site/docs/2.4.3/api/R/intersectAll.html            |  120 ++
 site/docs/2.4.3/api/R/isActive.html                |   71 +
 site/docs/2.4.3/api/R/isLocal.html                 |  108 ++
 site/docs/2.4.3/api/R/isStreaming.html             |  118 ++
 site/docs/2.4.3/api/R/join.html                    |  132 ++
 site/docs/2.4.3/api/R/last.html                    |   84 +
 site/docs/2.4.3/api/R/lastProgress.html            |   65 +
 site/docs/2.4.3/api/R/limit.html                   |  117 ++
 site/docs/2.4.3/api/R/listColumns.html             |   66 +
 site/docs/2.4.3/api/R/listDatabases.html           |   50 +
 site/docs/2.4.3/api/R/listFunctions.html           |   61 +
 site/docs/2.4.3/api/R/listTables.html              |   68 +
 site/docs/2.4.3/api/R/localCheckpoint.html         |  117 ++
 site/docs/2.4.3/api/R/match.html                   |   65 +
 site/docs/2.4.3/api/R/merge.html                   |  178 ++
 site/docs/2.4.3/api/R/mutate.html                  |  136 ++
 site/docs/2.4.3/api/R/nafunctions.html             |  174 ++
 site/docs/2.4.3/api/R/ncol.html                    |  105 ++
 site/docs/2.4.3/api/R/not.html                     |   90 +
 site/docs/2.4.3/api/R/nrow.html                    |  110 ++
 site/docs/2.4.3/api/R/orderBy.html                 |   85 +
 site/docs/2.4.3/api/R/otherwise.html               |   57 +
 site/docs/2.4.3/api/R/over.html                    |   79 +
 site/docs/2.4.3/api/R/partitionBy.html             |   78 +
 site/docs/2.4.3/api/R/persist.html                 |  114 ++
 site/docs/2.4.3/api/R/pivot.html                   |   80 +
 site/docs/2.4.3/api/R/predict.html                 |   54 +
 site/docs/2.4.3/api/R/print.jobj.html              |   45 +
 site/docs/2.4.3/api/R/print.structField.html       |   45 +
 site/docs/2.4.3/api/R/print.structType.html        |   45 +
 site/docs/2.4.3/api/R/printSchema.html             |  107 ++
 site/docs/2.4.3/api/R/queryName.html               |   75 +
 site/docs/2.4.3/api/R/randomSplit.html             |  116 ++
 site/docs/2.4.3/api/R/rangeBetween.html            |   87 +
 site/docs/2.4.3/api/R/rbind.html                   |  129 ++
 site/docs/2.4.3/api/R/read.df.html                 |  106 ++
 site/docs/2.4.3/api/R/read.jdbc.html               |  105 ++
 site/docs/2.4.3/api/R/read.json.html               |   77 +
 site/docs/2.4.3/api/R/read.ml.html                 |   66 +
 site/docs/2.4.3/api/R/read.orc.html                |   49 +
 site/docs/2.4.3/api/R/read.parquet.html            |   55 +
 site/docs/2.4.3/api/R/read.stream.html             |   93 +
 site/docs/2.4.3/api/R/read.text.html               |   74 +
 site/docs/2.4.3/api/R/recoverPartitions.html       |   56 +
 site/docs/2.4.3/api/R/refreshByPath.html           |   56 +
 site/docs/2.4.3/api/R/refreshTable.html            |   65 +
 .../2.4.3/api/R/registerTempTable-deprecated.html  |   70 +
 site/docs/2.4.3/api/R/rename.html                  |  141 ++
 site/docs/2.4.3/api/R/repartition.html             |  139 ++
 site/docs/2.4.3/api/R/repartitionByRange.html      |  134 ++
 site/docs/2.4.3/api/R/rollup.html                  |  129 ++
 site/docs/2.4.3/api/R/rowsBetween.html             |   87 +
 site/docs/2.4.3/api/R/sample.html                  |  133 ++
 site/docs/2.4.3/api/R/sampleBy.html                |   86 +
 site/docs/2.4.3/api/R/saveAsTable.html             |  141 ++
 site/docs/2.4.3/api/R/schema.html                  |  108 ++
 site/docs/2.4.3/api/R/select.html                  |  159 ++
 site/docs/2.4.3/api/R/selectExpr.html              |  122 ++
 site/docs/2.4.3/api/R/setCheckpointDir.html        |   60 +
 site/docs/2.4.3/api/R/setCurrentDatabase.html      |   54 +
 site/docs/2.4.3/api/R/setJobDescription.html       |   53 +
 site/docs/2.4.3/api/R/setJobGroup.html             |   65 +
 site/docs/2.4.3/api/R/setLocalProperty.html        |   59 +
 site/docs/2.4.3/api/R/setLogLevel.html             |   53 +
 site/docs/2.4.3/api/R/show.html                    |  126 ++
 site/docs/2.4.3/api/R/showDF.html                  |  128 ++
 site/docs/2.4.3/api/R/spark.addFile.html           |   69 +
 site/docs/2.4.3/api/R/spark.als.html               |  204 +++
 site/docs/2.4.3/api/R/spark.bisectingKmeans.html   |  179 ++
 site/docs/2.4.3/api/R/spark.decisionTree.html      |  234 +++
 site/docs/2.4.3/api/R/spark.fpGrowth.html          |  182 ++
 site/docs/2.4.3/api/R/spark.gaussianMixture.html   |  156 ++
 site/docs/2.4.3/api/R/spark.gbt.html               |  257 +++
 site/docs/2.4.3/api/R/spark.getSparkFiles.html     |   59 +
 .../api/R/spark.getSparkFilesRootDirectory.html    |   49 +
 site/docs/2.4.3/api/R/spark.glm.html               |  234 +++
 site/docs/2.4.3/api/R/spark.isoreg.html            |  146 ++
 site/docs/2.4.3/api/R/spark.kmeans.html            |  168 ++
 site/docs/2.4.3/api/R/spark.kstest.html            |  130 ++
 site/docs/2.4.3/api/R/spark.lapply.html            |   95 +
 site/docs/2.4.3/api/R/spark.lda.html               |  247 +++
 site/docs/2.4.3/api/R/spark.logit.html             |  269 +++
 site/docs/2.4.3/api/R/spark.mlp.html               |  192 ++
 site/docs/2.4.3/api/R/spark.naiveBayes.html        |  152 ++
 site/docs/2.4.3/api/R/spark.randomForest.html      |  260 +++
 site/docs/2.4.3/api/R/spark.survreg.html           |  156 ++
 site/docs/2.4.3/api/R/spark.svmLinear.html         |  177 ++
 site/docs/2.4.3/api/R/sparkR.callJMethod.html      |   90 +
 site/docs/2.4.3/api/R/sparkR.callJStatic.html      |   88 +
 site/docs/2.4.3/api/R/sparkR.conf.html             |   68 +
 site/docs/2.4.3/api/R/sparkR.init-deprecated.html  |   92 +
 site/docs/2.4.3/api/R/sparkR.newJObject.html       |   86 +
 site/docs/2.4.3/api/R/sparkR.session.html          |  114 ++
 site/docs/2.4.3/api/R/sparkR.session.stop.html     |   39 +
 site/docs/2.4.3/api/R/sparkR.uiWebUrl.html         |   50 +
 site/docs/2.4.3/api/R/sparkR.version.html          |   50 +
 .../2.4.3/api/R/sparkRHive.init-deprecated.html    |   67 +
 .../2.4.3/api/R/sparkRSQL.init-deprecated.html     |   68 +
 site/docs/2.4.3/api/R/sql.html                     |   64 +
 site/docs/2.4.3/api/R/startsWith.html              |   56 +
 site/docs/2.4.3/api/R/status.html                  |   65 +
 site/docs/2.4.3/api/R/stopQuery.html               |   66 +
 site/docs/2.4.3/api/R/storageLevel.html            |  107 ++
 site/docs/2.4.3/api/R/str.html                     |  108 ++
 site/docs/2.4.3/api/R/structField.html             |   83 +
 site/docs/2.4.3/api/R/structType.html              |   83 +
 site/docs/2.4.3/api/R/subset.html                  |  172 ++
 site/docs/2.4.3/api/R/substr.html                  |   73 +
 site/docs/2.4.3/api/R/summarize.html               |  144 ++
 site/docs/2.4.3/api/R/summary.html                 |  152 ++
 site/docs/2.4.3/api/R/tableNames.html              |   61 +
 site/docs/2.4.3/api/R/tableToDF.html               |   67 +
 site/docs/2.4.3/api/R/tables.html                  |   67 +
 site/docs/2.4.3/api/R/take.html                    |  112 ++
 site/docs/2.4.3/api/R/toJSON.html                  |  119 ++
 site/docs/2.4.3/api/R/uncacheTable.html            |   65 +
 site/docs/2.4.3/api/R/union.html                   |  137 ++
 site/docs/2.4.3/api/R/unionByName.html             |  130 ++
 site/docs/2.4.3/api/R/unpersist.html               |  118 ++
 site/docs/2.4.3/api/R/windowOrderBy.html           |   71 +
 site/docs/2.4.3/api/R/windowPartitionBy.html       |   71 +
 site/docs/2.4.3/api/R/with.html                    |  119 ++
 site/docs/2.4.3/api/R/withColumn.html              |  133 ++
 site/docs/2.4.3/api/R/withWatermark.html           |  149 ++
 site/docs/2.4.3/api/R/write.df.html                |  163 ++
 site/docs/2.4.3/api/R/write.jdbc.html              |  150 ++
 site/docs/2.4.3/api/R/write.json.html              |  124 ++
 site/docs/2.4.3/api/R/write.ml.html                |   62 +
 site/docs/2.4.3/api/R/write.orc.html               |  123 ++
 site/docs/2.4.3/api/R/write.parquet.html           |  131 ++
 site/docs/2.4.3/api/R/write.stream.html            |  183 ++
 site/docs/2.4.3/api/R/write.text.html              |  124 ++
 site/documentation.html                            |    9 +-
 site/downloads.html                                |   12 +-
 site/examples.html                                 |    8 +-
 site/faq.html                                      |    8 +-
 site/graphx/index.html                             |    8 +-
 site/history.html                                  |    8 +-
 site/improvement-proposals.html                    |    8 +-
 site/index.html                                    |    8 +-
 site/js/downloads.js                               |    8 +-
 site/mailing-lists.html                            |    8 +-
 site/mllib/index.html                              |    8 +-
 site/news/amp-camp-2013-registration-ope.html      |    8 +-
 site/news/announcing-the-first-spark-summit.html   |    8 +-
 site/news/fourth-spark-screencast-published.html   |    8 +-
 site/news/index.html                               |   17 +-
 site/news/nsdi-paper.html                          |    8 +-
 site/news/one-month-to-spark-summit-2015.html      |    8 +-
 .../news/proposals-open-for-spark-summit-east.html |    8 +-
 .../registration-open-for-spark-summit-east.html   |    8 +-
 site/news/run-spark-and-shark-on-amazon-emr.html   |    8 +-
 site/news/spark-0-6-1-and-0-5-2-released.html      |    8 +-
 site/news/spark-0-6-2-released.html                |    8 +-
 site/news/spark-0-7-0-released.html                |    8 +-
 site/news/spark-0-7-2-released.html                |    8 +-
 site/news/spark-0-7-3-released.html                |    8 +-
 site/news/spark-0-8-0-released.html                |    8 +-
 site/news/spark-0-8-1-released.html                |    8 +-
 site/news/spark-0-9-0-released.html                |    8 +-
 site/news/spark-0-9-1-released.html                |    8 +-
 site/news/spark-0-9-2-released.html                |    8 +-
 site/news/spark-1-0-0-released.html                |    8 +-
 site/news/spark-1-0-1-released.html                |    8 +-
 site/news/spark-1-0-2-released.html                |    8 +-
 site/news/spark-1-1-0-released.html                |    8 +-
 site/news/spark-1-1-1-released.html                |    8 +-
 site/news/spark-1-2-0-released.html                |    8 +-
 site/news/spark-1-2-1-released.html                |    8 +-
 site/news/spark-1-2-2-released.html                |    8 +-
 site/news/spark-1-3-0-released.html                |    8 +-
 site/news/spark-1-4-0-released.html                |    8 +-
 site/news/spark-1-4-1-released.html                |    8 +-
 site/news/spark-1-5-0-released.html                |    8 +-
 site/news/spark-1-5-1-released.html                |    8 +-
 site/news/spark-1-5-2-released.html                |    8 +-
 site/news/spark-1-6-0-released.html                |    8 +-
 site/news/spark-1-6-1-released.html                |    8 +-
 site/news/spark-1-6-2-released.html                |    8 +-
 site/news/spark-1-6-3-released.html                |    8 +-
 site/news/spark-2-0-0-released.html                |    8 +-
 site/news/spark-2-0-1-released.html                |    8 +-
 site/news/spark-2-0-2-released.html                |    8 +-
 site/news/spark-2-1-0-released.html                |    8 +-
 site/news/spark-2-1-1-released.html                |    8 +-
 site/news/spark-2-1-2-released.html                |    8 +-
 site/news/spark-2-1-3-released.html                |    8 +-
 site/news/spark-2-2-0-released.html                |    8 +-
 site/news/spark-2-2-1-released.html                |    8 +-
 site/news/spark-2-2-2-released.html                |    8 +-
 site/news/spark-2-3-0-released.html                |    8 +-
 site/news/spark-2-3-1-released.html                |    8 +-
 site/news/spark-2-3-2-released.html                |    8 +-
 site/news/spark-2-3-3-released.html                |    8 +-
 site/news/spark-2-4-0-released.html                |    8 +-
 site/news/spark-2-4-1-released.html                |    8 +-
 site/news/spark-2-4-2-released.html                |    8 +-
 ...1-0-released.html => spark-2-4-3-released.html} |   14 +-
 site/news/spark-2.0.0-preview.html                 |    8 +-
 .../news/spark-accepted-into-apache-incubator.html |    8 +-
 .../spark-ai-summit-apr-2019-agenda-posted.html    |    8 +-
 site/news/spark-and-shark-in-the-news.html         |    8 +-
 site/news/spark-becomes-tlp.html                   |    8 +-
 site/news/spark-featured-in-wired.html             |    8 +-
 .../news/spark-mailing-lists-moving-to-apache.html |    8 +-
 site/news/spark-meetups.html                       |    8 +-
 site/news/spark-release-2-2-3.html                 |    8 +-
 site/news/spark-screencasts-published.html         |    8 +-
 site/news/spark-summit-2013-is-a-wrap.html         |    8 +-
 site/news/spark-summit-2014-videos-posted.html     |    8 +-
 site/news/spark-summit-2015-videos-posted.html     |    8 +-
 site/news/spark-summit-agenda-posted.html          |    8 +-
 .../news/spark-summit-east-2015-videos-posted.html |    8 +-
 site/news/spark-summit-east-2016-cfp-closing.html  |    8 +-
 .../news/spark-summit-east-2017-agenda-posted.html |    8 +-
 site/news/spark-summit-east-agenda-posted.html     |    8 +-
 site/news/spark-summit-eu-2017-agenda-posted.html  |    8 +-
 site/news/spark-summit-europe-agenda-posted.html   |    8 +-
 site/news/spark-summit-europe.html                 |    8 +-
 .../news/spark-summit-june-2016-agenda-posted.html |    8 +-
 .../news/spark-summit-june-2017-agenda-posted.html |    8 +-
 .../news/spark-summit-june-2018-agenda-posted.html |    8 +-
 site/news/spark-summit-oct-2018-agenda-posted.html |    8 +-
 site/news/spark-tips-from-quantifind.html          |    8 +-
 .../spark-user-survey-and-powered-by-page.html     |    8 +-
 site/news/spark-version-0-6-0-released.html        |    8 +-
 .../news/spark-wins-cloudsort-100tb-benchmark.html |    8 +-
 ...ark-wins-daytona-gray-sort-100tb-benchmark.html |    8 +-
 .../strata-exercises-now-available-online.html     |    8 +-
 site/news/submit-talks-to-spark-summit-2014.html   |    8 +-
 site/news/submit-talks-to-spark-summit-2016.html   |    8 +-
 .../submit-talks-to-spark-summit-east-2016.html    |    8 +-
 .../news/submit-talks-to-spark-summit-eu-2016.html |    8 +-
 site/news/two-weeks-to-spark-summit-2014.html      |    8 +-
 .../video-from-first-spark-development-meetup.html |    8 +-
 site/powered-by.html                               |    8 +-
 site/release-process.html                          |    8 +-
 site/releases/spark-release-0-3.html               |    8 +-
 site/releases/spark-release-0-5-0.html             |    8 +-
 site/releases/spark-release-0-5-1.html             |    8 +-
 site/releases/spark-release-0-5-2.html             |    8 +-
 site/releases/spark-release-0-6-0.html             |    8 +-
 site/releases/spark-release-0-6-1.html             |    8 +-
 site/releases/spark-release-0-6-2.html             |    8 +-
 site/releases/spark-release-0-7-0.html             |    8 +-
 site/releases/spark-release-0-7-2.html             |    8 +-
 site/releases/spark-release-0-7-3.html             |    8 +-
 site/releases/spark-release-0-8-0.html             |    8 +-
 site/releases/spark-release-0-8-1.html             |    8 +-
 site/releases/spark-release-0-9-0.html             |    8 +-
 site/releases/spark-release-0-9-1.html             |    8 +-
 site/releases/spark-release-0-9-2.html             |    8 +-
 site/releases/spark-release-1-0-0.html             |    8 +-
 site/releases/spark-release-1-0-1.html             |    8 +-
 site/releases/spark-release-1-0-2.html             |    8 +-
 site/releases/spark-release-1-1-0.html             |    8 +-
 site/releases/spark-release-1-1-1.html             |    8 +-
 site/releases/spark-release-1-2-0.html             |    8 +-
 site/releases/spark-release-1-2-1.html             |    8 +-
 site/releases/spark-release-1-2-2.html             |    8 +-
 site/releases/spark-release-1-3-0.html             |    8 +-
 site/releases/spark-release-1-3-1.html             |    8 +-
 site/releases/spark-release-1-4-0.html             |    8 +-
 site/releases/spark-release-1-4-1.html             |    8 +-
 site/releases/spark-release-1-5-0.html             |    8 +-
 site/releases/spark-release-1-5-1.html             |    8 +-
 site/releases/spark-release-1-5-2.html             |    8 +-
 site/releases/spark-release-1-6-0.html             |    8 +-
 site/releases/spark-release-1-6-1.html             |    8 +-
 site/releases/spark-release-1-6-2.html             |    8 +-
 site/releases/spark-release-1-6-3.html             |    8 +-
 site/releases/spark-release-2-0-0.html             |    8 +-
 site/releases/spark-release-2-0-1.html             |    8 +-
 site/releases/spark-release-2-0-2.html             |    8 +-
 site/releases/spark-release-2-1-0.html             |    8 +-
 site/releases/spark-release-2-1-1.html             |    8 +-
 site/releases/spark-release-2-1-2.html             |    8 +-
 site/releases/spark-release-2-1-3.html             |    8 +-
 site/releases/spark-release-2-2-0.html             |    8 +-
 site/releases/spark-release-2-2-1.html             |    8 +-
 site/releases/spark-release-2-2-2.html             |    8 +-
 site/releases/spark-release-2-2-3.html             |    8 +-
 site/releases/spark-release-2-3-0.html             |    8 +-
 site/releases/spark-release-2-3-1.html             |    8 +-
 site/releases/spark-release-2-3-2.html             |    8 +-
 site/releases/spark-release-2-3-3.html             |    8 +-
 site/releases/spark-release-2-4-0.html             |    8 +-
 site/releases/spark-release-2-4-1.html             |    8 +-
 site/releases/spark-release-2-4-2.html             |    8 +-
 ...release-2-4-2.html => spark-release-2-4-3.html} |   24 +-
 site/research.html                                 |    8 +-
 site/screencasts/1-first-steps-with-spark.html     |    8 +-
 .../2-spark-documentation-overview.html            |    8 +-
 .../screencasts/3-transformations-and-caching.html |    8 +-
 site/screencasts/4-a-standalone-job-in-spark.html  |    8 +-
 site/screencasts/index.html                        |    8 +-
 site/security.html                                 |    8 +-
 site/sitemap.xml                                   |   18 +-
 site/sql/index.html                                |    8 +-
 site/streaming/index.html                          |    8 +-
 site/third-party-projects.html                     |    8 +-
 site/trademarks.html                               |    8 +-
 site/versioning-policy.html                        |    8 +-
 413 files changed, 27877 insertions(+), 723 deletions(-)

diff --git a/_layouts/global.html b/_layouts/global.html
index 27d705e..9ff8e30 100644
--- a/_layouts/global.html
+++ b/_layouts/global.html
@@ -121,7 +121,7 @@
           Documentation <b class="caret"></b>
         </a>
         <ul class="dropdown-menu">
-          <li><a href="{{site.baseurl}}/docs/latest/">Latest Release (Spark 2.4.2)</a></li>
+          <li><a href="{{site.baseurl}}/docs/latest/">Latest Release (Spark 2.4.3)</a></li>
           <li><a href="{{site.baseurl}}/documentation.html">Older Versions and Other Resources</a></li>
           <li><a href="{{site.baseurl}}/faq.html">Frequently Asked Questions</a></li>
         </ul>
diff --git a/documentation.md b/documentation.md
index f9ae29e..f4581cc 100644
--- a/documentation.md
+++ b/documentation.md
@@ -12,6 +12,7 @@ navigation:
 <p>Setup instructions, programming guides, and other documentation are available for each stable version of Spark below:</p>
 
 <ul>
+  <li><a href="{{site.baseurl}}/docs/2.4.3/">Spark 2.4.3</a></li>
   <li><a href="{{site.baseurl}}/docs/2.4.2/">Spark 2.4.2</a></li>
   <li><a href="{{site.baseurl}}/docs/2.4.1/">Spark 2.4.1</a></li>
   <li><a href="{{site.baseurl}}/docs/2.4.0/">Spark 2.4.0</a></li>
diff --git a/downloads.md b/downloads.md
index 57e4a06..4c9fed9 100644
--- a/downloads.md
+++ b/downloads.md
@@ -26,14 +26,14 @@ $(document).ready(function() {
 
 4. Verify this release using the <span id="sparkDownloadVerify"></span> and [project release KEYS](https://www.apache.org/dist/spark/KEYS).
 
-Note that, Spark is pre-built with Scala 2.12 since version 2.4.2. Previous versions are pre-built with Scala 2.11.
+Note that, Spark is pre-built with Scala 2.11 except version 2.4.2, which is pre-built with Scala 2.12.
 
 ### Link with Spark
 Spark artifacts are [hosted in Maven Central](https://search.maven.org/search?q=g:org.apache.spark). You can add a Maven dependency with the following coordinates:
 
     groupId: org.apache.spark
     artifactId: spark-core_2.11
-    version: 2.4.2
+    version: 2.4.3
 
 ### Installing with PyPi
 <a href="https://pypi.org/project/pyspark/">PySpark</a> is now available in pypi. To install just run `pip install pyspark`.
diff --git a/js/downloads.js b/js/downloads.js
index 1373494..bf56d75 100644
--- a/js/downloads.js
+++ b/js/downloads.js
@@ -15,14 +15,14 @@ var sources = {pretty: "Source Code", tag: "sources"};
 var hadoopFree = {pretty: "Pre-build with user-provided Apache Hadoop", tag: "without-hadoop"};
 var hadoop2p6 = {pretty: "Pre-built for Apache Hadoop 2.6", tag: "hadoop2.6"};
 var hadoop2p7 = {pretty: "Pre-built for Apache Hadoop 2.7 and later", tag: "hadoop2.7"};
-var scala2p11_hadoopFree = {pretty: "Pre-build with Scala 2.11 and user-provided Apache Hadoop", tag: "without-hadoop-scala-2.11"};
+var scala2p12_hadoopFree = {pretty: "Pre-build with Scala 2.12 and user-provided Apache Hadoop", tag: "without-hadoop-scala-2.12"};
 
 // 2.2.0+
 var packagesV8 = [hadoop2p7, hadoop2p6, hadoopFree, sources];
-// 2.4.2+
-var packagesV9 = [hadoop2p7, hadoop2p6, hadoopFree, scala2p11_hadoopFree, sources];
+// 2.4.0+
+var packagesV9 = [hadoop2p7, hadoop2p6, hadoopFree, scala2p12_hadoopFree, sources];
 
-addRelease("2.4.2", new Date("04/23/2019"), packagesV9, true);
+addRelease("2.4.3", new Date("05/07/2019"), packagesV9, true);
 addRelease("2.3.3", new Date("02/15/2019"), packagesV8, true);
 
 function append(el, contents) {
diff --git a/news/_posts/2019-05-08-spark-2-4-3-released.md b/news/_posts/2019-05-08-spark-2-4-3-released.md
new file mode 100644
index 0000000..4731409
--- /dev/null
+++ b/news/_posts/2019-05-08-spark-2-4-3-released.md
@@ -0,0 +1,14 @@
+---
+layout: post
+title: Spark 2.4.3 released
+categories:
+- News
+tags: []
+status: publish
+type: post
+published: true
+meta:
+  _edit_last: '4'
+  _wpas_done_all: '1'
+---
+We are happy to announce the availability of <a href="{{site.baseurl}}/releases/spark-release-2-4-3.html" title="Spark Release 2.4.3">Spark 2.4.3</a>! Visit the <a href="{{site.baseurl}}/releases/spark-release-2-4-3.html" title="Spark Release 2.4.3">release notes</a> to read about the new features, or <a href="{{site.baseurl}}/downloads.html">download</a> the release today.
\ No newline at end of file
diff --git a/releases/_posts/2019-05-08-spark-release-2-4-3.md b/releases/_posts/2019-05-08-spark-release-2-4-3.md
new file mode 100644
index 0000000..0d04399
--- /dev/null
+++ b/releases/_posts/2019-05-08-spark-release-2-4-3.md
@@ -0,0 +1,24 @@
+---
+layout: post
+title: Spark Release 2.4.3
+categories: []
+tags: []
+status: publish
+type: post
+published: true
+meta:
+  _edit_last: '4'
+  _wpas_done_all: '1'
+---
+
+Spark 2.4.3 is a maintenance release containing stability fixes. This release is based on the branch-2.4 maintenance branch of Spark. We strongly recommend all 2.4 users to upgrade to this stable release.
+
+Note that 2.4.3 switched the default Scala version from Scala 2.12 to Scala 2.11, which is the default for all the previous 2.x releases except 2.4.2. That means, the pre-built convenience binaries are compiled for Scala 2.11.
+Spark is still cross-published for 2.11 and 2.12 in Maven Central, and can be built for 2.12 from source.
+
+### Notable changes
+  - [[SPARK-27494]](https://issues.apache.org/jira/browse/SPARK-27494): It fixes a correctness bug in Kafka source v2, when processing NULL in the key value.
+
+You can consult JIRA for the [detailed changes](https://s.apache.org/spark-2.4.3).
+
+We would like to acknowledge all community members for contributing patches to this release.
\ No newline at end of file
diff --git a/site/committers.html b/site/committers.html
index faed196..bcc7c21 100644
--- a/site/committers.html
+++ b/site/committers.html
@@ -106,7 +106,7 @@
           Documentation <b class="caret"></b>
         </a>
         <ul class="dropdown-menu">
-          <li><a href="/docs/latest/">Latest Release (Spark 2.4.2)</a></li>
+          <li><a href="/docs/latest/">Latest Release (Spark 2.4.3)</a></li>
           <li><a href="/documentation.html">Older Versions and Other Resources</a></li>
           <li><a href="/faq.html">Frequently Asked Questions</a></li>
         </ul>
@@ -162,6 +162,9 @@
       <h5>Latest News</h5>
       <ul class="list-unstyled">
         
+          <li><a href="/news/spark-2-4-3-released.html">Spark 2.4.3 released</a>
+          <span class="small">(May 08, 2019)</span></li>
+        
           <li><a href="/news/spark-2-4-2-released.html">Spark 2.4.2 released</a>
           <span class="small">(Apr 23, 2019)</span></li>
         
@@ -171,9 +174,6 @@
           <li><a href="/news/spark-2-3-3-released.html">Spark 2.3.3 released</a>
           <span class="small">(Feb 15, 2019)</span></li>
         
-          <li><a href="/news/spark-release-2-2-3.html">Spark 2.2.3 released</a>
-          <span class="small">(Jan 11, 2019)</span></li>
-        
       </ul>
       <p class="small" style="text-align: right;"><a href="/news/index.html">Archive</a></p>
     </div>
diff --git a/site/community.html b/site/community.html
index c8ef734..fb55f25 100644
--- a/site/community.html
+++ b/site/community.html
@@ -106,7 +106,7 @@
           Documentation <b class="caret"></b>
         </a>
         <ul class="dropdown-menu">
-          <li><a href="/docs/latest/">Latest Release (Spark 2.4.2)</a></li>
+          <li><a href="/docs/latest/">Latest Release (Spark 2.4.3)</a></li>
           <li><a href="/documentation.html">Older Versions and Other Resources</a></li>
           <li><a href="/faq.html">Frequently Asked Questions</a></li>
         </ul>
@@ -162,6 +162,9 @@
       <h5>Latest News</h5>
       <ul class="list-unstyled">
         
+          <li><a href="/news/spark-2-4-3-released.html">Spark 2.4.3 released</a>
+          <span class="small">(May 08, 2019)</span></li>
+        
           <li><a href="/news/spark-2-4-2-released.html">Spark 2.4.2 released</a>
           <span class="small">(Apr 23, 2019)</span></li>
         
@@ -171,9 +174,6 @@
           <li><a href="/news/spark-2-3-3-released.html">Spark 2.3.3 released</a>
           <span class="small">(Feb 15, 2019)</span></li>
         
-          <li><a href="/news/spark-release-2-2-3.html">Spark 2.2.3 released</a>
-          <span class="small">(Jan 11, 2019)</span></li>
-        
       </ul>
       <p class="small" style="text-align: right;"><a href="/news/index.html">Archive</a></p>
     </div>
diff --git a/site/contributing.html b/site/contributing.html
index 70496f1..3d76ab7 100644
--- a/site/contributing.html
+++ b/site/contributing.html
@@ -106,7 +106,7 @@
           Documentation <b class="caret"></b>
         </a>
         <ul class="dropdown-menu">
-          <li><a href="/docs/latest/">Latest Release (Spark 2.4.2)</a></li>
+          <li><a href="/docs/latest/">Latest Release (Spark 2.4.3)</a></li>
           <li><a href="/documentation.html">Older Versions and Other Resources</a></li>
           <li><a href="/faq.html">Frequently Asked Questions</a></li>
         </ul>
@@ -162,6 +162,9 @@
       <h5>Latest News</h5>
       <ul class="list-unstyled">
         
+          <li><a href="/news/spark-2-4-3-released.html">Spark 2.4.3 released</a>
+          <span class="small">(May 08, 2019)</span></li>
+        
           <li><a href="/news/spark-2-4-2-released.html">Spark 2.4.2 released</a>
           <span class="small">(Apr 23, 2019)</span></li>
         
@@ -171,9 +174,6 @@
           <li><a href="/news/spark-2-3-3-released.html">Spark 2.3.3 released</a>
           <span class="small">(Feb 15, 2019)</span></li>
         
-          <li><a href="/news/spark-release-2-2-3.html">Spark 2.2.3 released</a>
-          <span class="small">(Jan 11, 2019)</span></li>
-        
       </ul>
       <p class="small" style="text-align: right;"><a href="/news/index.html">Archive</a></p>
     </div>
diff --git a/site/developer-tools.html b/site/developer-tools.html
index 8a41c8e..75055f4 100644
--- a/site/developer-tools.html
+++ b/site/developer-tools.html
@@ -106,7 +106,7 @@
           Documentation <b class="caret"></b>
         </a>
         <ul class="dropdown-menu">
-          <li><a href="/docs/latest/">Latest Release (Spark 2.4.2)</a></li>
+          <li><a href="/docs/latest/">Latest Release (Spark 2.4.3)</a></li>
           <li><a href="/documentation.html">Older Versions and Other Resources</a></li>
           <li><a href="/faq.html">Frequently Asked Questions</a></li>
         </ul>
@@ -162,6 +162,9 @@
       <h5>Latest News</h5>
       <ul class="list-unstyled">
         
+          <li><a href="/news/spark-2-4-3-released.html">Spark 2.4.3 released</a>
+          <span class="small">(May 08, 2019)</span></li>
+        
           <li><a href="/news/spark-2-4-2-released.html">Spark 2.4.2 released</a>
           <span class="small">(Apr 23, 2019)</span></li>
         
@@ -171,9 +174,6 @@
           <li><a href="/news/spark-2-3-3-released.html">Spark 2.3.3 released</a>
           <span class="small">(Feb 15, 2019)</span></li>
         
-          <li><a href="/news/spark-release-2-2-3.html">Spark 2.2.3 released</a>
-          <span class="small">(Jan 11, 2019)</span></li>
-        
       </ul>
       <p class="small" style="text-align: right;"><a href="/news/index.html">Archive</a></p>
     </div>
diff --git a/site/docs/2.4.3/api/R/00Index.html b/site/docs/2.4.3/api/R/00Index.html
index 3d07dea..095e572 100644
--- a/site/docs/2.4.3/api/R/00Index.html
+++ b/site/docs/2.4.3/api/R/00Index.html
@@ -19,5 +19,1931 @@
 <h2>Help Pages</h2>
 
 
-There are no help pages in this package
+<p style="text-align: center;">
+<a href="#A">A</a>
+<a href="#B">B</a>
+<a href="#C">C</a>
+<a href="#D">D</a>
+<a href="#E">E</a>
+<a href="#F">F</a>
+<a href="#G">G</a>
+<a href="#H">H</a>
+<a href="#I">I</a>
+<a href="#J">J</a>
+<a href="#K">K</a>
+<a href="#L">L</a>
+<a href="#M">M</a>
+<a href="#N">N</a>
+<a href="#O">O</a>
+<a href="#P">P</a>
+<a href="#Q">Q</a>
+<a href="#R">R</a>
+<a href="#S">S</a>
+<a href="#T">T</a>
+<a href="#U">U</a>
+<a href="#V">V</a>
+<a href="#W">W</a>
+<a href="#Y">Y</a>
+<a href="#misc">misc</a>
+</p>
+
+
+<h2><a name="A">-- A --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_math_functions.html">abs</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">abs-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">acos</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">acos-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">add_months</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">add_months-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="AFTSurvivalRegressionModel-class.html">AFTSurvivalRegressionModel-class</a></td>
+<td>S4 class that represents a AFTSurvivalRegressionModel</td></tr>
+<tr><td style="width: 25%;"><a href="summarize.html">agg</a></td>
+<td>summarize</td></tr>
+<tr><td style="width: 25%;"><a href="summarize.html">agg-method</a></td>
+<td>summarize</td></tr>
+<tr><td style="width: 25%;"><a href="alias.html">alias</a></td>
+<td>alias</td></tr>
+<tr><td style="width: 25%;"><a href="alias.html">alias-method</a></td>
+<td>alias</td></tr>
+<tr><td style="width: 25%;"><a href="ALSModel-class.html">ALSModel-class</a></td>
+<td>S4 class that represents an ALSModel</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">approxCountDistinct</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">approxCountDistinct-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="approxQuantile.html">approxQuantile</a></td>
+<td>Calculates the approximate quantiles of numerical columns of a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="approxQuantile.html">approxQuantile-method</a></td>
+<td>Calculates the approximate quantiles of numerical columns of a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="arrange.html">arrange</a></td>
+<td>Arrange Rows by Variables</td></tr>
+<tr><td style="width: 25%;"><a href="arrange.html">arrange-method</a></td>
+<td>Arrange Rows by Variables</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">arrays_overlap</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">arrays_overlap-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">arrays_zip</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">arrays_zip-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_contains</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_contains-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_distinct</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_distinct-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_except</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_except-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_intersect</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_intersect-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_join</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_join-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_max</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_max-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_min</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_min-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_position</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_position-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_remove</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_remove-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_repeat</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_repeat-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_sort</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_sort-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_union</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">array_union-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="as.data.frame.html">as.data.frame</a></td>
+<td>Download data from a SparkDataFrame into a R data.frame</td></tr>
+<tr><td style="width: 25%;"><a href="as.data.frame.html">as.data.frame-method</a></td>
+<td>Download data from a SparkDataFrame into a R data.frame</td></tr>
+<tr><td style="width: 25%;"><a href="createDataFrame.html">as.DataFrame</a></td>
+<td>Create a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="createDataFrame.html">as.DataFrame.default</a></td>
+<td>Create a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">asc</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">ascii</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">ascii-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">asin</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">asin-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">associationRules-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">atan</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">atan-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">atan2</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">atan2-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="attach.html">attach</a></td>
+<td>Attach SparkDataFrame to R search path</td></tr>
+<tr><td style="width: 25%;"><a href="attach.html">attach-method</a></td>
+<td>Attach SparkDataFrame to R search path</td></tr>
+<tr><td style="width: 25%;"><a href="avg.html">avg</a></td>
+<td>avg</td></tr>
+<tr><td style="width: 25%;"><a href="avg.html">avg-method</a></td>
+<td>avg</td></tr>
+<tr><td style="width: 25%;"><a href="awaitTermination.html">awaitTermination</a></td>
+<td>awaitTermination</td></tr>
+<tr><td style="width: 25%;"><a href="awaitTermination.html">awaitTermination-method</a></td>
+<td>awaitTermination</td></tr>
+</table>
+
+<h2><a name="B">-- B --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_string_functions.html">base64</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">base64-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="between.html">between</a></td>
+<td>between</td></tr>
+<tr><td style="width: 25%;"><a href="between.html">between-method</a></td>
+<td>between</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">bin</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">bin-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="BisectingKMeansModel-class.html">BisectingKMeansModel-class</a></td>
+<td>S4 class that represents a BisectingKMeansModel</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">bitwiseNOT</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">bitwiseNOT-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="broadcast.html">broadcast</a></td>
+<td>broadcast</td></tr>
+<tr><td style="width: 25%;"><a href="broadcast.html">broadcast-method</a></td>
+<td>broadcast</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">bround</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">bround-method</a></td>
+<td>Math functions for Column operations</td></tr>
+</table>
+
+<h2><a name="C">-- C --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="cache.html">cache</a></td>
+<td>Cache</td></tr>
+<tr><td style="width: 25%;"><a href="cache.html">cache-method</a></td>
+<td>Cache</td></tr>
+<tr><td style="width: 25%;"><a href="cacheTable.html">cacheTable</a></td>
+<td>Cache Table</td></tr>
+<tr><td style="width: 25%;"><a href="cacheTable.html">cacheTable.default</a></td>
+<td>Cache Table</td></tr>
+<tr><td style="width: 25%;"><a href="cancelJobGroup.html">cancelJobGroup</a></td>
+<td>Cancel active jobs for the specified group</td></tr>
+<tr><td style="width: 25%;"><a href="cancelJobGroup.html">cancelJobGroup.default</a></td>
+<td>Cancel active jobs for the specified group</td></tr>
+<tr><td style="width: 25%;"><a href="cast.html">cast</a></td>
+<td>Casts the column to a different data type.</td></tr>
+<tr><td style="width: 25%;"><a href="cast.html">cast-method</a></td>
+<td>Casts the column to a different data type.</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">cbrt</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">cbrt-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">ceil</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">ceil-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">ceiling</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">ceiling-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="checkpoint.html">checkpoint</a></td>
+<td>checkpoint</td></tr>
+<tr><td style="width: 25%;"><a href="checkpoint.html">checkpoint-method</a></td>
+<td>checkpoint</td></tr>
+<tr><td style="width: 25%;"><a href="clearCache.html">clearCache</a></td>
+<td>Clear Cache</td></tr>
+<tr><td style="width: 25%;"><a href="clearCache.html">clearCache.default</a></td>
+<td>Clear Cache</td></tr>
+<tr><td style="width: 25%;"><a href="clearJobGroup.html">clearJobGroup</a></td>
+<td>Clear current job group ID and its description</td></tr>
+<tr><td style="width: 25%;"><a href="clearJobGroup.html">clearJobGroup.default</a></td>
+<td>Clear current job group ID and its description</td></tr>
+<tr><td style="width: 25%;"><a href="coalesce.html">coalesce</a></td>
+<td>Coalesce</td></tr>
+<tr><td style="width: 25%;"><a href="coalesce.html">coalesce-method</a></td>
+<td>Coalesce</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">coalesce-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="collect.html">collect</a></td>
+<td>Collects all the elements of a SparkDataFrame and coerces them into an R data.frame.</td></tr>
+<tr><td style="width: 25%;"><a href="collect.html">collect-method</a></td>
+<td>Collects all the elements of a SparkDataFrame and coerces them into an R data.frame.</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">collect_list</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">collect_list-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">collect_set</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">collect_set-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">colnames</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">colnames-method</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">colnames&lt;-</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">colnames&lt;--method</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="coltypes.html">coltypes</a></td>
+<td>coltypes</td></tr>
+<tr><td style="width: 25%;"><a href="coltypes.html">coltypes-method</a></td>
+<td>coltypes</td></tr>
+<tr><td style="width: 25%;"><a href="coltypes.html">coltypes&lt;-</a></td>
+<td>coltypes</td></tr>
+<tr><td style="width: 25%;"><a href="coltypes.html">coltypes&lt;--method</a></td>
+<td>coltypes</td></tr>
+<tr><td style="width: 25%;"><a href="column.html">column</a></td>
+<td>S4 class that represents a SparkDataFrame column</td></tr>
+<tr><td style="width: 25%;"><a href="column.html">Column-class</a></td>
+<td>S4 class that represents a SparkDataFrame column</td></tr>
+<tr><td style="width: 25%;"><a href="column.html">column-method</a></td>
+<td>S4 class that represents a SparkDataFrame column</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">columnfunctions</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">columns</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">columns-method</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">column_aggregate_functions</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">column_collection_functions</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">column_datetime_diff_functions</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">column_datetime_functions</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">column_math_functions</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">column_misc_functions</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">column_nonaggregate_functions</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">column_string_functions</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">column_window_functions</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">concat</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">concat-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">concat_ws</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">concat_ws-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">contains</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">conv</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">conv-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="corr.html">corr</a></td>
+<td>corr</td></tr>
+<tr><td style="width: 25%;"><a href="corr.html">corr-method</a></td>
+<td>corr</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">cos</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">cos-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">cosh</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">cosh-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="count.html">count</a></td>
+<td>Count</td></tr>
+<tr><td style="width: 25%;"><a href="count.html">count-method</a></td>
+<td>Count</td></tr>
+<tr><td style="width: 25%;"><a href="nrow.html">count-method</a></td>
+<td>Returns the number of rows in a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">countDistinct</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">countDistinct-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="cov.html">cov</a></td>
+<td>cov</td></tr>
+<tr><td style="width: 25%;"><a href="cov.html">cov-method</a></td>
+<td>cov</td></tr>
+<tr><td style="width: 25%;"><a href="cov.html">covar_pop</a></td>
+<td>cov</td></tr>
+<tr><td style="width: 25%;"><a href="cov.html">covar_pop-method</a></td>
+<td>cov</td></tr>
+<tr><td style="width: 25%;"><a href="cov.html">covar_samp</a></td>
+<td>cov</td></tr>
+<tr><td style="width: 25%;"><a href="cov.html">covar_samp-method</a></td>
+<td>cov</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">crc32</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">crc32-method</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="createDataFrame.html">createDataFrame</a></td>
+<td>Create a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="createDataFrame.html">createDataFrame.default</a></td>
+<td>Create a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="createExternalTable-deprecated.html">createExternalTable</a></td>
+<td>(Deprecated) Create an external table</td></tr>
+<tr><td style="width: 25%;"><a href="createExternalTable-deprecated.html">createExternalTable.default</a></td>
+<td>(Deprecated) Create an external table</td></tr>
+<tr><td style="width: 25%;"><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></td>
+<td>Creates a temporary view using the given name.</td></tr>
+<tr><td style="width: 25%;"><a href="createOrReplaceTempView.html">createOrReplaceTempView-method</a></td>
+<td>Creates a temporary view using the given name.</td></tr>
+<tr><td style="width: 25%;"><a href="createTable.html">createTable</a></td>
+<td>Creates a table based on the dataset in a data source</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">create_array</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">create_array-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">create_map</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">create_map-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="crossJoin.html">crossJoin</a></td>
+<td>CrossJoin</td></tr>
+<tr><td style="width: 25%;"><a href="crossJoin.html">crossJoin-method</a></td>
+<td>CrossJoin</td></tr>
+<tr><td style="width: 25%;"><a href="crosstab.html">crosstab</a></td>
+<td>Computes a pair-wise frequency table of the given columns</td></tr>
+<tr><td style="width: 25%;"><a href="crosstab.html">crosstab-method</a></td>
+<td>Computes a pair-wise frequency table of the given columns</td></tr>
+<tr><td style="width: 25%;"><a href="cube.html">cube</a></td>
+<td>cube</td></tr>
+<tr><td style="width: 25%;"><a href="cube.html">cube-method</a></td>
+<td>cube</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">cume_dist</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">cume_dist-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="currentDatabase.html">currentDatabase</a></td>
+<td>Returns the current default database</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">current_date</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">current_date-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">current_timestamp</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">current_timestamp-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+</table>
+
+<h2><a name="D">-- D --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="dapply.html">dapply</a></td>
+<td>dapply</td></tr>
+<tr><td style="width: 25%;"><a href="dapply.html">dapply-method</a></td>
+<td>dapply</td></tr>
+<tr><td style="width: 25%;"><a href="dapplyCollect.html">dapplyCollect</a></td>
+<td>dapplyCollect</td></tr>
+<tr><td style="width: 25%;"><a href="dapplyCollect.html">dapplyCollect-method</a></td>
+<td>dapplyCollect</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">datediff</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">datediff-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">date_add</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">date_add-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">date_format</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">date_format-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">date_sub</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">date_sub-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">date_trunc</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">date_trunc-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">dayofmonth</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">dayofmonth-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">dayofweek</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">dayofweek-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">dayofyear</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">dayofyear-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="DecisionTreeClassificationModel-class.html">DecisionTreeClassificationModel-class</a></td>
+<td>S4 class that represents a DecisionTreeClassificationModel</td></tr>
+<tr><td style="width: 25%;"><a href="DecisionTreeRegressionModel-class.html">DecisionTreeRegressionModel-class</a></td>
+<td>S4 class that represents a DecisionTreeRegressionModel</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">decode</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">decode-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">dense_rank</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">dense_rank-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">desc</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="describe.html">describe</a></td>
+<td>describe</td></tr>
+<tr><td style="width: 25%;"><a href="describe.html">describe-method</a></td>
+<td>describe</td></tr>
+<tr><td style="width: 25%;"><a href="dim.html">dim</a></td>
+<td>Returns the dimensions of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="dim.html">dim-method</a></td>
+<td>Returns the dimensions of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="distinct.html">distinct</a></td>
+<td>Distinct</td></tr>
+<tr><td style="width: 25%;"><a href="distinct.html">distinct-method</a></td>
+<td>Distinct</td></tr>
+<tr><td style="width: 25%;"><a href="drop.html">drop</a></td>
+<td>drop</td></tr>
+<tr><td style="width: 25%;"><a href="drop.html">drop-method</a></td>
+<td>drop</td></tr>
+<tr><td style="width: 25%;"><a href="dropDuplicates.html">dropDuplicates</a></td>
+<td>dropDuplicates</td></tr>
+<tr><td style="width: 25%;"><a href="dropDuplicates.html">dropDuplicates-method</a></td>
+<td>dropDuplicates</td></tr>
+<tr><td style="width: 25%;"><a href="nafunctions.html">dropna</a></td>
+<td>A set of SparkDataFrame functions working with NA values</td></tr>
+<tr><td style="width: 25%;"><a href="nafunctions.html">dropna-method</a></td>
+<td>A set of SparkDataFrame functions working with NA values</td></tr>
+<tr><td style="width: 25%;"><a href="dropTempTable-deprecated.html">dropTempTable</a></td>
+<td>(Deprecated) Drop Temporary Table</td></tr>
+<tr><td style="width: 25%;"><a href="dropTempTable-deprecated.html">dropTempTable.default</a></td>
+<td>(Deprecated) Drop Temporary Table</td></tr>
+<tr><td style="width: 25%;"><a href="dropTempView.html">dropTempView</a></td>
+<td>Drops the temporary view with the given view name in the catalog.</td></tr>
+<tr><td style="width: 25%;"><a href="dtypes.html">dtypes</a></td>
+<td>DataTypes</td></tr>
+<tr><td style="width: 25%;"><a href="dtypes.html">dtypes-method</a></td>
+<td>DataTypes</td></tr>
+</table>
+
+<h2><a name="E">-- E --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">element_at</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">element_at-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">encode</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">encode-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="endsWith.html">endsWith</a></td>
+<td>endsWith</td></tr>
+<tr><td style="width: 25%;"><a href="endsWith.html">endsWith-method</a></td>
+<td>endsWith</td></tr>
+<tr><td style="width: 25%;"><a href="except.html">except</a></td>
+<td>except</td></tr>
+<tr><td style="width: 25%;"><a href="except.html">except-method</a></td>
+<td>except</td></tr>
+<tr><td style="width: 25%;"><a href="exceptAll.html">exceptAll</a></td>
+<td>exceptAll</td></tr>
+<tr><td style="width: 25%;"><a href="exceptAll.html">exceptAll-method</a></td>
+<td>exceptAll</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">exp</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">exp-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="explain.html">explain</a></td>
+<td>Explain</td></tr>
+<tr><td style="width: 25%;"><a href="explain.html">explain-method</a></td>
+<td>Explain</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">explode</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">explode-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">explode_outer</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">explode_outer-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">expm1</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">expm1-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">expr</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">expr-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+</table>
+
+<h2><a name="F">-- F --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_math_functions.html">factorial</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">factorial-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="nafunctions.html">fillna</a></td>
+<td>A set of SparkDataFrame functions working with NA values</td></tr>
+<tr><td style="width: 25%;"><a href="nafunctions.html">fillna-method</a></td>
+<td>A set of SparkDataFrame functions working with NA values</td></tr>
+<tr><td style="width: 25%;"><a href="filter.html">filter</a></td>
+<td>Filter</td></tr>
+<tr><td style="width: 25%;"><a href="filter.html">filter-method</a></td>
+<td>Filter</td></tr>
+<tr><td style="width: 25%;"><a href="first.html">first</a></td>
+<td>Return the first row of a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="first.html">first-method</a></td>
+<td>Return the first row of a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="fitted.html">fitted</a></td>
+<td>Get fitted result from a k-means model</td></tr>
+<tr><td style="width: 25%;"><a href="fitted.html">fitted-method</a></td>
+<td>Get fitted result from a k-means model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.bisectingKmeans.html">fitted-method</a></td>
+<td>Bisecting K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">flatten</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">flatten-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">floor</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">floor-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">format_number</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">format_number-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">format_string</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">format_string-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="FPGrowthModel-class.html">FPGrowthModel-class</a></td>
+<td>S4 class that represents a FPGrowthModel</td></tr>
+<tr><td style="width: 25%;"><a href="freqItems.html">freqItems</a></td>
+<td>Finding frequent items for columns, possibly with false positives</td></tr>
+<tr><td style="width: 25%;"><a href="freqItems.html">freqItems-method</a></td>
+<td>Finding frequent items for columns, possibly with false positives</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">freqItemsets-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">from_json</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">from_json-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">from_unixtime</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">from_unixtime-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">from_utc_timestamp</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">from_utc_timestamp-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+</table>
+
+<h2><a name="G">-- G --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="gapply.html">gapply</a></td>
+<td>gapply</td></tr>
+<tr><td style="width: 25%;"><a href="gapply.html">gapply-method</a></td>
+<td>gapply</td></tr>
+<tr><td style="width: 25%;"><a href="gapplyCollect.html">gapplyCollect</a></td>
+<td>gapplyCollect</td></tr>
+<tr><td style="width: 25%;"><a href="gapplyCollect.html">gapplyCollect-method</a></td>
+<td>gapplyCollect</td></tr>
+<tr><td style="width: 25%;"><a href="GaussianMixtureModel-class.html">GaussianMixtureModel-class</a></td>
+<td>S4 class that represents a GaussianMixtureModel</td></tr>
+<tr><td style="width: 25%;"><a href="GBTClassificationModel-class.html">GBTClassificationModel-class</a></td>
+<td>S4 class that represents a GBTClassificationModel</td></tr>
+<tr><td style="width: 25%;"><a href="GBTRegressionModel-class.html">GBTRegressionModel-class</a></td>
+<td>S4 class that represents a GBTRegressionModel</td></tr>
+<tr><td style="width: 25%;"><a href="GeneralizedLinearRegressionModel-class.html">GeneralizedLinearRegressionModel-class</a></td>
+<td>S4 class that represents a generalized linear model</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">getField</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">getItem</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="getLocalProperty.html">getLocalProperty</a></td>
+<td>Get a local property set in this thread, or 'NULL' if it is missing. See 'setLocalProperty'.</td></tr>
+<tr><td style="width: 25%;"><a href="getNumPartitions.html">getNumPartitions</a></td>
+<td>getNumPartitions</td></tr>
+<tr><td style="width: 25%;"><a href="getNumPartitions.html">getNumPartitions-method</a></td>
+<td>getNumPartitions</td></tr>
+<tr><td style="width: 25%;"><a href="glm.html">glm</a></td>
+<td>Generalized Linear Models (R-compliant)</td></tr>
+<tr><td style="width: 25%;"><a href="glm.html">glm-method</a></td>
+<td>Generalized Linear Models (R-compliant)</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">greatest</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">greatest-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="groupBy.html">groupBy</a></td>
+<td>GroupBy</td></tr>
+<tr><td style="width: 25%;"><a href="groupBy.html">groupBy-method</a></td>
+<td>GroupBy</td></tr>
+<tr><td style="width: 25%;"><a href="GroupedData.html">groupedData</a></td>
+<td>S4 class that represents a GroupedData</td></tr>
+<tr><td style="width: 25%;"><a href="GroupedData.html">GroupedData-class</a></td>
+<td>S4 class that represents a GroupedData</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">grouping_bit</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">grouping_bit-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">grouping_id</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">grouping_id-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="groupBy.html">group_by</a></td>
+<td>GroupBy</td></tr>
+<tr><td style="width: 25%;"><a href="groupBy.html">group_by-method</a></td>
+<td>GroupBy</td></tr>
+</table>
+
+<h2><a name="H">-- H --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">hash</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">hash-method</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="hashCode.html">hashCode</a></td>
+<td>Compute the hashCode of an object</td></tr>
+<tr><td style="width: 25%;"><a href="head.html">head</a></td>
+<td>Head</td></tr>
+<tr><td style="width: 25%;"><a href="head.html">head-method</a></td>
+<td>Head</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">hex</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">hex-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="hint.html">hint</a></td>
+<td>hint</td></tr>
+<tr><td style="width: 25%;"><a href="hint.html">hint-method</a></td>
+<td>hint</td></tr>
+<tr><td style="width: 25%;"><a href="histogram.html">histogram</a></td>
+<td>Compute histogram statistics for given column</td></tr>
+<tr><td style="width: 25%;"><a href="histogram.html">histogram-method</a></td>
+<td>Compute histogram statistics for given column</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">hour</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">hour-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">hypot</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">hypot-method</a></td>
+<td>Math functions for Column operations</td></tr>
+</table>
+
+<h2><a name="I">-- I --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">ifelse</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">ifelse-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">initcap</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">initcap-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">input_file_name</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">input_file_name-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="insertInto.html">insertInto</a></td>
+<td>insertInto</td></tr>
+<tr><td style="width: 25%;"><a href="insertInto.html">insertInto-method</a></td>
+<td>insertInto</td></tr>
+<tr><td style="width: 25%;"><a href="install.spark.html">install.spark</a></td>
+<td>Download and Install Apache Spark to a Local Directory</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">instr</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">instr-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="intersect.html">intersect</a></td>
+<td>Intersect</td></tr>
+<tr><td style="width: 25%;"><a href="intersect.html">intersect-method</a></td>
+<td>Intersect</td></tr>
+<tr><td style="width: 25%;"><a href="intersectAll.html">intersectAll</a></td>
+<td>intersectAll</td></tr>
+<tr><td style="width: 25%;"><a href="intersectAll.html">intersectAll-method</a></td>
+<td>intersectAll</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">is.nan</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">is.nan-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="isActive.html">isActive</a></td>
+<td>isActive</td></tr>
+<tr><td style="width: 25%;"><a href="isActive.html">isActive-method</a></td>
+<td>isActive</td></tr>
+<tr><td style="width: 25%;"><a href="isLocal.html">isLocal</a></td>
+<td>isLocal</td></tr>
+<tr><td style="width: 25%;"><a href="isLocal.html">isLocal-method</a></td>
+<td>isLocal</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">isNaN</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">isnan</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">isnan-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">isNotNull</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">isNull</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="IsotonicRegressionModel-class.html">IsotonicRegressionModel-class</a></td>
+<td>S4 class that represents an IsotonicRegressionModel</td></tr>
+<tr><td style="width: 25%;"><a href="isStreaming.html">isStreaming</a></td>
+<td>isStreaming</td></tr>
+<tr><td style="width: 25%;"><a href="isStreaming.html">isStreaming-method</a></td>
+<td>isStreaming</td></tr>
+</table>
+
+<h2><a name="J">-- J --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="join.html">join</a></td>
+<td>Join</td></tr>
+<tr><td style="width: 25%;"><a href="join.html">join-method</a></td>
+<td>Join</td></tr>
+<tr><td style="width: 25%;"><a href="read.json.html">jsonFile</a></td>
+<td>Create a SparkDataFrame from a JSON file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.json.html">jsonFile.default</a></td>
+<td>Create a SparkDataFrame from a JSON file.</td></tr>
+</table>
+
+<h2><a name="K">-- K --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="KMeansModel-class.html">KMeansModel-class</a></td>
+<td>S4 class that represents a KMeansModel</td></tr>
+<tr><td style="width: 25%;"><a href="KSTest-class.html">KSTest-class</a></td>
+<td>S4 class that represents an KSTest</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">kurtosis</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">kurtosis-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+</table>
+
+<h2><a name="L">-- L --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_window_functions.html">lag</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">lag-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="last.html">last</a></td>
+<td>last</td></tr>
+<tr><td style="width: 25%;"><a href="last.html">last-method</a></td>
+<td>last</td></tr>
+<tr><td style="width: 25%;"><a href="lastProgress.html">lastProgress</a></td>
+<td>lastProgress</td></tr>
+<tr><td style="width: 25%;"><a href="lastProgress.html">lastProgress-method</a></td>
+<td>lastProgress</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">last_day</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">last_day-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="LDAModel-class.html">LDAModel-class</a></td>
+<td>S4 class that represents an LDAModel</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">lead</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">lead-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">least</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">least-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">length</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">length-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">levenshtein</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">levenshtein-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">like</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="limit.html">limit</a></td>
+<td>Limit</td></tr>
+<tr><td style="width: 25%;"><a href="limit.html">limit-method</a></td>
+<td>Limit</td></tr>
+<tr><td style="width: 25%;"><a href="LinearSVCModel-class.html">LinearSVCModel-class</a></td>
+<td>S4 class that represents an LinearSVCModel</td></tr>
+<tr><td style="width: 25%;"><a href="listColumns.html">listColumns</a></td>
+<td>Returns a list of columns for the given table/view in the specified database</td></tr>
+<tr><td style="width: 25%;"><a href="listDatabases.html">listDatabases</a></td>
+<td>Returns a list of databases available</td></tr>
+<tr><td style="width: 25%;"><a href="listFunctions.html">listFunctions</a></td>
+<td>Returns a list of functions registered in the specified database</td></tr>
+<tr><td style="width: 25%;"><a href="listTables.html">listTables</a></td>
+<td>Returns a list of tables or views in the specified database</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">lit</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">lit-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="read.df.html">loadDF</a></td>
+<td>Load a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="read.df.html">loadDF.default</a></td>
+<td>Load a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="localCheckpoint.html">localCheckpoint</a></td>
+<td>localCheckpoint</td></tr>
+<tr><td style="width: 25%;"><a href="localCheckpoint.html">localCheckpoint-method</a></td>
+<td>localCheckpoint</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">locate</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">locate-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log10</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log10-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log1p</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log1p-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log2</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">log2-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="LogisticRegressionModel-class.html">LogisticRegressionModel-class</a></td>
+<td>S4 class that represents an LogisticRegressionModel</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">lower</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">lower-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">lpad</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">lpad-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">ltrim</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">ltrim-method</a></td>
+<td>String functions for Column operations</td></tr>
+</table>
+
+<h2><a name="M">-- M --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">map_from_arrays</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">map_from_arrays-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">map_keys</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">map_keys-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">map_values</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">map_values-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">max</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">max-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">md5</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">md5-method</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">mean</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">mean-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="merge.html">merge</a></td>
+<td>Merges two data frames</td></tr>
+<tr><td style="width: 25%;"><a href="merge.html">merge-method</a></td>
+<td>Merges two data frames</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">min</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">min-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">minute</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">minute-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">monotonically_increasing_id</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">monotonically_increasing_id-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">month</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">month-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">months_between</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">months_between-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="MultilayerPerceptronClassificationModel-class.html">MultilayerPerceptronClassificationModel-class</a></td>
+<td>S4 class that represents a MultilayerPerceptronClassificationModel</td></tr>
+<tr><td style="width: 25%;"><a href="mutate.html">mutate</a></td>
+<td>Mutate</td></tr>
+<tr><td style="width: 25%;"><a href="mutate.html">mutate-method</a></td>
+<td>Mutate</td></tr>
+</table>
+
+<h2><a name="N">-- N --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="count.html">n</a></td>
+<td>Count</td></tr>
+<tr><td style="width: 25%;"><a href="count.html">n-method</a></td>
+<td>Count</td></tr>
+<tr><td style="width: 25%;"><a href="nafunctions.html">na.omit</a></td>
+<td>A set of SparkDataFrame functions working with NA values</td></tr>
+<tr><td style="width: 25%;"><a href="nafunctions.html">na.omit-method</a></td>
+<td>A set of SparkDataFrame functions working with NA values</td></tr>
+<tr><td style="width: 25%;"><a href="NaiveBayesModel-class.html">NaiveBayesModel-class</a></td>
+<td>S4 class that represents a NaiveBayesModel</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">names</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">names-method</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">names&lt;-</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="columns.html">names&lt;--method</a></td>
+<td>Column Names of SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">nanvl</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">nanvl-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="ncol.html">ncol</a></td>
+<td>Returns the number of columns in a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="ncol.html">ncol-method</a></td>
+<td>Returns the number of columns in a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">negate</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">negate-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">next_day</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">next_day-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="not.html">not</a></td>
+<td>!</td></tr>
+<tr><td style="width: 25%;"><a href="not.html">not-method</a></td>
+<td>!</td></tr>
+<tr><td style="width: 25%;"><a href="nrow.html">nrow</a></td>
+<td>Returns the number of rows in a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="nrow.html">nrow-method</a></td>
+<td>Returns the number of rows in a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">ntile</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">ntile-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">n_distinct</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">n_distinct-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+</table>
+
+<h2><a name="O">-- O --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="orderBy.html">orderBy</a></td>
+<td>Ordering Columns in a WindowSpec</td></tr>
+<tr><td style="width: 25%;"><a href="arrange.html">orderBy-method</a></td>
+<td>Arrange Rows by Variables</td></tr>
+<tr><td style="width: 25%;"><a href="orderBy.html">orderBy-method</a></td>
+<td>Ordering Columns in a WindowSpec</td></tr>
+<tr><td style="width: 25%;"><a href="otherwise.html">otherwise</a></td>
+<td>otherwise</td></tr>
+<tr><td style="width: 25%;"><a href="otherwise.html">otherwise-method</a></td>
+<td>otherwise</td></tr>
+<tr><td style="width: 25%;"><a href="over.html">over</a></td>
+<td>over</td></tr>
+<tr><td style="width: 25%;"><a href="over.html">over-method</a></td>
+<td>over</td></tr>
+</table>
+
+<h2><a name="P">-- P --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="read.parquet.html">parquetFile</a></td>
+<td>Create a SparkDataFrame from a Parquet file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.parquet.html">parquetFile.default</a></td>
+<td>Create a SparkDataFrame from a Parquet file.</td></tr>
+<tr><td style="width: 25%;"><a href="partitionBy.html">partitionBy</a></td>
+<td>partitionBy</td></tr>
+<tr><td style="width: 25%;"><a href="partitionBy.html">partitionBy-method</a></td>
+<td>partitionBy</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">percent_rank</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">percent_rank-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="persist.html">persist</a></td>
+<td>Persist</td></tr>
+<tr><td style="width: 25%;"><a href="persist.html">persist-method</a></td>
+<td>Persist</td></tr>
+<tr><td style="width: 25%;"><a href="pivot.html">pivot</a></td>
+<td>Pivot a column of the GroupedData and perform the specified aggregation.</td></tr>
+<tr><td style="width: 25%;"><a href="pivot.html">pivot-method</a></td>
+<td>Pivot a column of the GroupedData and perform the specified aggregation.</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">pmod</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">pmod-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">posexplode</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">posexplode-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">posexplode_outer</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">posexplode_outer-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="predict.html">predict</a></td>
+<td>Makes predictions from a MLlib model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.als.html">predict-method</a></td>
+<td>Alternating Least Squares (ALS) for Collaborative Filtering</td></tr>
+<tr><td style="width: 25%;"><a href="spark.bisectingKmeans.html">predict-method</a></td>
+<td>Bisecting K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">predict-method</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">predict-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gaussianMixture.html">predict-method</a></td>
+<td>Multivariate Gaussian Mixture Model (GMM)</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">predict-method</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.glm.html">predict-method</a></td>
+<td>Generalized Linear Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.isoreg.html">predict-method</a></td>
+<td>Isotonic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kmeans.html">predict-method</a></td>
+<td>K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.logit.html">predict-method</a></td>
+<td>Logistic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.mlp.html">predict-method</a></td>
+<td>Multilayer Perceptron Classification Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.naiveBayes.html">predict-method</a></td>
+<td>Naive Bayes Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">predict-method</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.survreg.html">predict-method</a></td>
+<td>Accelerated Failure Time (AFT) Survival Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.svmLinear.html">predict-method</a></td>
+<td>Linear SVM Model</td></tr>
+<tr><td style="width: 25%;"><a href="print.jobj.html">print.jobj</a></td>
+<td>Print a JVM object reference.</td></tr>
+<tr><td style="width: 25%;"><a href="print.structField.html">print.structField</a></td>
+<td>Print a Spark StructField.</td></tr>
+<tr><td style="width: 25%;"><a href="print.structType.html">print.structType</a></td>
+<td>Print a Spark StructType.</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">print.summary.DecisionTreeClassificationModel</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">print.summary.DecisionTreeRegressionModel</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">print.summary.GBTClassificationModel</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">print.summary.GBTRegressionModel</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.glm.html">print.summary.GeneralizedLinearRegressionModel</a></td>
+<td>Generalized Linear Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kstest.html">print.summary.KSTest</a></td>
+<td>(One-Sample) Kolmogorov-Smirnov Test</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">print.summary.RandomForestClassificationModel</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">print.summary.RandomForestRegressionModel</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="printSchema.html">printSchema</a></td>
+<td>Print Schema of a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="printSchema.html">printSchema-method</a></td>
+<td>Print Schema of a SparkDataFrame</td></tr>
+</table>
+
+<h2><a name="Q">-- Q --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">quarter</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">quarter-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="queryName.html">queryName</a></td>
+<td>queryName</td></tr>
+<tr><td style="width: 25%;"><a href="queryName.html">queryName-method</a></td>
+<td>queryName</td></tr>
+</table>
+
+<h2><a name="R">-- R --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">rand</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">rand-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">randn</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">randn-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="RandomForestClassificationModel-class.html">RandomForestClassificationModel-class</a></td>
+<td>S4 class that represents a RandomForestClassificationModel</td></tr>
+<tr><td style="width: 25%;"><a href="RandomForestRegressionModel-class.html">RandomForestRegressionModel-class</a></td>
+<td>S4 class that represents a RandomForestRegressionModel</td></tr>
+<tr><td style="width: 25%;"><a href="randomSplit.html">randomSplit</a></td>
+<td>randomSplit</td></tr>
+<tr><td style="width: 25%;"><a href="randomSplit.html">randomSplit-method</a></td>
+<td>randomSplit</td></tr>
+<tr><td style="width: 25%;"><a href="rangeBetween.html">rangeBetween</a></td>
+<td>rangeBetween</td></tr>
+<tr><td style="width: 25%;"><a href="rangeBetween.html">rangeBetween-method</a></td>
+<td>rangeBetween</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">rank</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">rank-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="rbind.html">rbind</a></td>
+<td>Union two or more SparkDataFrames</td></tr>
+<tr><td style="width: 25%;"><a href="rbind.html">rbind-method</a></td>
+<td>Union two or more SparkDataFrames</td></tr>
+<tr><td style="width: 25%;"><a href="read.df.html">read.df</a></td>
+<td>Load a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="read.df.html">read.df.default</a></td>
+<td>Load a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="read.jdbc.html">read.jdbc</a></td>
+<td>Create a SparkDataFrame representing the database table accessible via JDBC URL</td></tr>
+<tr><td style="width: 25%;"><a href="read.json.html">read.json</a></td>
+<td>Create a SparkDataFrame from a JSON file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.json.html">read.json.default</a></td>
+<td>Create a SparkDataFrame from a JSON file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.ml.html">read.ml</a></td>
+<td>Load a fitted MLlib model from the input path.</td></tr>
+<tr><td style="width: 25%;"><a href="read.orc.html">read.orc</a></td>
+<td>Create a SparkDataFrame from an ORC file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.parquet.html">read.parquet</a></td>
+<td>Create a SparkDataFrame from a Parquet file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.parquet.html">read.parquet.default</a></td>
+<td>Create a SparkDataFrame from a Parquet file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.stream.html">read.stream</a></td>
+<td>Load a streaming SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="read.text.html">read.text</a></td>
+<td>Create a SparkDataFrame from a text file.</td></tr>
+<tr><td style="width: 25%;"><a href="read.text.html">read.text.default</a></td>
+<td>Create a SparkDataFrame from a text file.</td></tr>
+<tr><td style="width: 25%;"><a href="recoverPartitions.html">recoverPartitions</a></td>
+<td>Recovers all the partitions in the directory of a table and update the catalog</td></tr>
+<tr><td style="width: 25%;"><a href="refreshByPath.html">refreshByPath</a></td>
+<td>Invalidates and refreshes all the cached data and metadata for SparkDataFrame containing path</td></tr>
+<tr><td style="width: 25%;"><a href="refreshTable.html">refreshTable</a></td>
+<td>Invalidates and refreshes all the cached data and metadata of the given table</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">regexp_extract</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">regexp_extract-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">regexp_replace</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">regexp_replace-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="registerTempTable-deprecated.html">registerTempTable</a></td>
+<td>(Deprecated) Register Temporary Table</td></tr>
+<tr><td style="width: 25%;"><a href="registerTempTable-deprecated.html">registerTempTable-method</a></td>
+<td>(Deprecated) Register Temporary Table</td></tr>
+<tr><td style="width: 25%;"><a href="rename.html">rename</a></td>
+<td>rename</td></tr>
+<tr><td style="width: 25%;"><a href="rename.html">rename-method</a></td>
+<td>rename</td></tr>
+<tr><td style="width: 25%;"><a href="repartition.html">repartition</a></td>
+<td>Repartition</td></tr>
+<tr><td style="width: 25%;"><a href="repartition.html">repartition-method</a></td>
+<td>Repartition</td></tr>
+<tr><td style="width: 25%;"><a href="repartitionByRange.html">repartitionByRange</a></td>
+<td>Repartition by range</td></tr>
+<tr><td style="width: 25%;"><a href="repartitionByRange.html">repartitionByRange-method</a></td>
+<td>Repartition by range</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">repeat_string</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">repeat_string-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">reverse</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">reverse-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">rint</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">rint-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="columnfunctions.html">rlike</a></td>
+<td>A set of operations working with SparkDataFrame columns</td></tr>
+<tr><td style="width: 25%;"><a href="rollup.html">rollup</a></td>
+<td>rollup</td></tr>
+<tr><td style="width: 25%;"><a href="rollup.html">rollup-method</a></td>
+<td>rollup</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">round</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">round-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="rowsBetween.html">rowsBetween</a></td>
+<td>rowsBetween</td></tr>
+<tr><td style="width: 25%;"><a href="rowsBetween.html">rowsBetween-method</a></td>
+<td>rowsBetween</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">row_number</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_window_functions.html">row_number-method</a></td>
+<td>Window functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">rpad</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">rpad-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">rtrim</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">rtrim-method</a></td>
+<td>String functions for Column operations</td></tr>
+</table>
+
+<h2><a name="S">-- S --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="sample.html">sample</a></td>
+<td>Sample</td></tr>
+<tr><td style="width: 25%;"><a href="sample.html">sample-method</a></td>
+<td>Sample</td></tr>
+<tr><td style="width: 25%;"><a href="sampleBy.html">sampleBy</a></td>
+<td>Returns a stratified sample without replacement</td></tr>
+<tr><td style="width: 25%;"><a href="sampleBy.html">sampleBy-method</a></td>
+<td>Returns a stratified sample without replacement</td></tr>
+<tr><td style="width: 25%;"><a href="sample.html">sample_frac</a></td>
+<td>Sample</td></tr>
+<tr><td style="width: 25%;"><a href="sample.html">sample_frac-method</a></td>
+<td>Sample</td></tr>
+<tr><td style="width: 25%;"><a href="write.parquet.html">saveAsParquetFile</a></td>
+<td>Save the contents of SparkDataFrame as a Parquet file, preserving the schema.</td></tr>
+<tr><td style="width: 25%;"><a href="write.parquet.html">saveAsParquetFile-method</a></td>
+<td>Save the contents of SparkDataFrame as a Parquet file, preserving the schema.</td></tr>
+<tr><td style="width: 25%;"><a href="saveAsTable.html">saveAsTable</a></td>
+<td>Save the contents of the SparkDataFrame to a data source as a table</td></tr>
+<tr><td style="width: 25%;"><a href="saveAsTable.html">saveAsTable-method</a></td>
+<td>Save the contents of the SparkDataFrame to a data source as a table</td></tr>
+<tr><td style="width: 25%;"><a href="write.df.html">saveDF</a></td>
+<td>Save the contents of SparkDataFrame to a data source.</td></tr>
+<tr><td style="width: 25%;"><a href="write.df.html">saveDF-method</a></td>
+<td>Save the contents of SparkDataFrame to a data source.</td></tr>
+<tr><td style="width: 25%;"><a href="schema.html">schema</a></td>
+<td>Get schema object</td></tr>
+<tr><td style="width: 25%;"><a href="schema.html">schema-method</a></td>
+<td>Get schema object</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">sd</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">sd-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">second</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">second-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="select.html">select</a></td>
+<td>Select</td></tr>
+<tr><td style="width: 25%;"><a href="select.html">select-method</a></td>
+<td>Select</td></tr>
+<tr><td style="width: 25%;"><a href="selectExpr.html">selectExpr</a></td>
+<td>SelectExpr</td></tr>
+<tr><td style="width: 25%;"><a href="selectExpr.html">selectExpr-method</a></td>
+<td>SelectExpr</td></tr>
+<tr><td style="width: 25%;"><a href="setCheckpointDir.html">setCheckpointDir</a></td>
+<td>Set checkpoint directory</td></tr>
+<tr><td style="width: 25%;"><a href="setCurrentDatabase.html">setCurrentDatabase</a></td>
+<td>Sets the current default database</td></tr>
+<tr><td style="width: 25%;"><a href="setJobDescription.html">setJobDescription</a></td>
+<td>Set a human readable description of the current job.</td></tr>
+<tr><td style="width: 25%;"><a href="setJobGroup.html">setJobGroup</a></td>
+<td>Assigns a group ID to all the jobs started by this thread until the group ID is set to a different value or cleared.</td></tr>
+<tr><td style="width: 25%;"><a href="setJobGroup.html">setJobGroup.default</a></td>
+<td>Assigns a group ID to all the jobs started by this thread until the group ID is set to a different value or cleared.</td></tr>
+<tr><td style="width: 25%;"><a href="setLocalProperty.html">setLocalProperty</a></td>
+<td>Set a local property that affects jobs submitted from this thread, such as the Spark fair scheduler pool.</td></tr>
+<tr><td style="width: 25%;"><a href="setLogLevel.html">setLogLevel</a></td>
+<td>Set new log level</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">sha1</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">sha1-method</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">sha2</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_misc_functions.html">sha2-method</a></td>
+<td>Miscellaneous functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">shiftLeft</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">shiftLeft-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">shiftRight</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">shiftRight-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">shiftRightUnsigned</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">shiftRightUnsigned-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="show.html">show</a></td>
+<td>show</td></tr>
+<tr><td style="width: 25%;"><a href="show.html">show-method</a></td>
+<td>show</td></tr>
+<tr><td style="width: 25%;"><a href="showDF.html">showDF</a></td>
+<td>showDF</td></tr>
+<tr><td style="width: 25%;"><a href="showDF.html">showDF-method</a></td>
+<td>showDF</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">shuffle</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">shuffle-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sign</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sign-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">signum</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">signum-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sin</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sin-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sinh</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sinh-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">size</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">size-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">skewness</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">skewness-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">slice</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">slice-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">sort_array</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">sort_array-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">soundex</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">soundex-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="spark.addFile.html">spark.addFile</a></td>
+<td>Add a file or directory to be downloaded with this Spark job on every node.</td></tr>
+<tr><td style="width: 25%;"><a href="spark.als.html">spark.als</a></td>
+<td>Alternating Least Squares (ALS) for Collaborative Filtering</td></tr>
+<tr><td style="width: 25%;"><a href="spark.als.html">spark.als-method</a></td>
+<td>Alternating Least Squares (ALS) for Collaborative Filtering</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">spark.associationRules</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">spark.associationRules-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.bisectingKmeans.html">spark.bisectingKmeans</a></td>
+<td>Bisecting K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.bisectingKmeans.html">spark.bisectingKmeans-method</a></td>
+<td>Bisecting K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">spark.decisionTree</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">spark.decisionTree-method</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">spark.fpGrowth</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">spark.fpGrowth-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">spark.freqItemsets</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">spark.freqItemsets-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gaussianMixture.html">spark.gaussianMixture</a></td>
+<td>Multivariate Gaussian Mixture Model (GMM)</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gaussianMixture.html">spark.gaussianMixture-method</a></td>
+<td>Multivariate Gaussian Mixture Model (GMM)</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">spark.gbt</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">spark.gbt-method</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.getSparkFiles.html">spark.getSparkFiles</a></td>
+<td>Get the absolute path of a file added through spark.addFile.</td></tr>
+<tr><td style="width: 25%;"><a href="spark.getSparkFilesRootDirectory.html">spark.getSparkFilesRootDirectory</a></td>
+<td>Get the root directory that contains files added through spark.addFile.</td></tr>
+<tr><td style="width: 25%;"><a href="spark.glm.html">spark.glm</a></td>
+<td>Generalized Linear Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.glm.html">spark.glm-method</a></td>
+<td>Generalized Linear Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.isoreg.html">spark.isoreg</a></td>
+<td>Isotonic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.isoreg.html">spark.isoreg-method</a></td>
+<td>Isotonic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kmeans.html">spark.kmeans</a></td>
+<td>K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kmeans.html">spark.kmeans-method</a></td>
+<td>K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kstest.html">spark.kstest</a></td>
+<td>(One-Sample) Kolmogorov-Smirnov Test</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kstest.html">spark.kstest-method</a></td>
+<td>(One-Sample) Kolmogorov-Smirnov Test</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lapply.html">spark.lapply</a></td>
+<td>Run a function over a list of elements, distributing the computations with Spark</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">spark.lda</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">spark.lda-method</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.logit.html">spark.logit</a></td>
+<td>Logistic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.logit.html">spark.logit-method</a></td>
+<td>Logistic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.mlp.html">spark.mlp</a></td>
+<td>Multilayer Perceptron Classification Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.mlp.html">spark.mlp-method</a></td>
+<td>Multilayer Perceptron Classification Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.naiveBayes.html">spark.naiveBayes</a></td>
+<td>Naive Bayes Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.naiveBayes.html">spark.naiveBayes-method</a></td>
+<td>Naive Bayes Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">spark.perplexity</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">spark.perplexity-method</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">spark.posterior</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">spark.posterior-method</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">spark.randomForest</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">spark.randomForest-method</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.survreg.html">spark.survreg</a></td>
+<td>Accelerated Failure Time (AFT) Survival Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.survreg.html">spark.survreg-method</a></td>
+<td>Accelerated Failure Time (AFT) Survival Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.svmLinear.html">spark.svmLinear</a></td>
+<td>Linear SVM Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.svmLinear.html">spark.svmLinear-method</a></td>
+<td>Linear SVM Model</td></tr>
+<tr><td style="width: 25%;"><a href="SparkDataFrame.html">SparkDataFrame-class</a></td>
+<td>S4 class that represents a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.callJMethod.html">sparkR.callJMethod</a></td>
+<td>Call Java Methods</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.callJStatic.html">sparkR.callJStatic</a></td>
+<td>Call Static Java Methods</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.conf.html">sparkR.conf</a></td>
+<td>Get Runtime Config from the current active SparkSession</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.init-deprecated.html">sparkR.init</a></td>
+<td>(Deprecated) Initialize a new Spark Context</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.newJObject.html">sparkR.newJObject</a></td>
+<td>Create Java Objects</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.session.html">sparkR.session</a></td>
+<td>Get the existing SparkSession or initialize a new SparkSession.</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.session.stop.html">sparkR.session.stop</a></td>
+<td>Stop the Spark Session and Spark Context</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.session.stop.html">sparkR.stop</a></td>
+<td>Stop the Spark Session and Spark Context</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.uiWebUrl.html">sparkR.uiWebUrl</a></td>
+<td>Get the URL of the SparkUI instance for the current active SparkSession</td></tr>
+<tr><td style="width: 25%;"><a href="sparkR.version.html">sparkR.version</a></td>
+<td>Get version of Spark on which this application is running</td></tr>
+<tr><td style="width: 25%;"><a href="sparkRHive.init-deprecated.html">sparkRHive.init</a></td>
+<td>(Deprecated) Initialize a new HiveContext</td></tr>
+<tr><td style="width: 25%;"><a href="sparkRSQL.init-deprecated.html">sparkRSQL.init</a></td>
+<td>(Deprecated) Initialize a new SQLContext</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">spark_partition_id</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">spark_partition_id-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">split_string</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">split_string-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="sql.html">sql</a></td>
+<td>SQL Query</td></tr>
+<tr><td style="width: 25%;"><a href="sql.html">sql.default</a></td>
+<td>SQL Query</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sqrt</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">sqrt-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="startsWith.html">startsWith</a></td>
+<td>startsWith</td></tr>
+<tr><td style="width: 25%;"><a href="startsWith.html">startsWith-method</a></td>
+<td>startsWith</td></tr>
+<tr><td style="width: 25%;"><a href="status.html">status</a></td>
+<td>status</td></tr>
+<tr><td style="width: 25%;"><a href="status.html">status-method</a></td>
+<td>status</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">stddev</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">stddev-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">stddev_pop</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">stddev_pop-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">stddev_samp</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">stddev_samp-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="stopQuery.html">stopQuery</a></td>
+<td>stopQuery</td></tr>
+<tr><td style="width: 25%;"><a href="stopQuery.html">stopQuery-method</a></td>
+<td>stopQuery</td></tr>
+<tr><td style="width: 25%;"><a href="storageLevel.html">storageLevel</a></td>
+<td>StorageLevel</td></tr>
+<tr><td style="width: 25%;"><a href="storageLevel.html">storageLevel-method</a></td>
+<td>StorageLevel</td></tr>
+<tr><td style="width: 25%;"><a href="str.html">str</a></td>
+<td>Compactly display the structure of a dataset</td></tr>
+<tr><td style="width: 25%;"><a href="str.html">str-method</a></td>
+<td>Compactly display the structure of a dataset</td></tr>
+<tr><td style="width: 25%;"><a href="StreamingQuery.html">StreamingQuery-class</a></td>
+<td>S4 class that represents a StreamingQuery</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">struct</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">struct-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="structField.html">structField</a></td>
+<td>structField</td></tr>
+<tr><td style="width: 25%;"><a href="structField.html">structField.character</a></td>
+<td>structField</td></tr>
+<tr><td style="width: 25%;"><a href="structField.html">structField.jobj</a></td>
+<td>structField</td></tr>
+<tr><td style="width: 25%;"><a href="structType.html">structType</a></td>
+<td>structType</td></tr>
+<tr><td style="width: 25%;"><a href="structType.html">structType.character</a></td>
+<td>structType</td></tr>
+<tr><td style="width: 25%;"><a href="structType.html">structType.jobj</a></td>
+<td>structType</td></tr>
+<tr><td style="width: 25%;"><a href="structType.html">structType.structField</a></td>
+<td>structType</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">subset</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">subset-method</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="substr.html">substr</a></td>
+<td>substr</td></tr>
+<tr><td style="width: 25%;"><a href="substr.html">substr-method</a></td>
+<td>substr</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">substring_index</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">substring_index-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">sum</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">sum-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">sumDistinct</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">sumDistinct-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="summarize.html">summarize</a></td>
+<td>summarize</td></tr>
+<tr><td style="width: 25%;"><a href="summarize.html">summarize-method</a></td>
+<td>summarize</td></tr>
+<tr><td style="width: 25%;"><a href="summary.html">summary</a></td>
+<td>summary</td></tr>
+<tr><td style="width: 25%;"><a href="spark.als.html">summary-method</a></td>
+<td>Alternating Least Squares (ALS) for Collaborative Filtering</td></tr>
+<tr><td style="width: 25%;"><a href="spark.bisectingKmeans.html">summary-method</a></td>
+<td>Bisecting K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">summary-method</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gaussianMixture.html">summary-method</a></td>
+<td>Multivariate Gaussian Mixture Model (GMM)</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">summary-method</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.glm.html">summary-method</a></td>
+<td>Generalized Linear Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.isoreg.html">summary-method</a></td>
+<td>Isotonic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kmeans.html">summary-method</a></td>
+<td>K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kstest.html">summary-method</a></td>
+<td>(One-Sample) Kolmogorov-Smirnov Test</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">summary-method</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.logit.html">summary-method</a></td>
+<td>Logistic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.mlp.html">summary-method</a></td>
+<td>Multilayer Perceptron Classification Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.naiveBayes.html">summary-method</a></td>
+<td>Naive Bayes Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">summary-method</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.survreg.html">summary-method</a></td>
+<td>Accelerated Failure Time (AFT) Survival Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.svmLinear.html">summary-method</a></td>
+<td>Linear SVM Model</td></tr>
+<tr><td style="width: 25%;"><a href="summary.html">summary-method</a></td>
+<td>summary</td></tr>
+</table>
+
+<h2><a name="T">-- T --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="tableNames.html">tableNames</a></td>
+<td>Table Names</td></tr>
+<tr><td style="width: 25%;"><a href="tableNames.html">tableNames.default</a></td>
+<td>Table Names</td></tr>
+<tr><td style="width: 25%;"><a href="tables.html">tables</a></td>
+<td>Tables</td></tr>
+<tr><td style="width: 25%;"><a href="tables.html">tables.default</a></td>
+<td>Tables</td></tr>
+<tr><td style="width: 25%;"><a href="tableToDF.html">tableToDF</a></td>
+<td>Create a SparkDataFrame from a SparkSQL table or view</td></tr>
+<tr><td style="width: 25%;"><a href="take.html">take</a></td>
+<td>Take the first NUM rows of a SparkDataFrame and return the results as a R data.frame</td></tr>
+<tr><td style="width: 25%;"><a href="take.html">take-method</a></td>
+<td>Take the first NUM rows of a SparkDataFrame and return the results as a R data.frame</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">tan</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">tan-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">tanh</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">tanh-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">toDegrees</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">toDegrees-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="toJSON.html">toJSON</a></td>
+<td>toJSON</td></tr>
+<tr><td style="width: 25%;"><a href="toJSON.html">toJSON-method</a></td>
+<td>toJSON</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">toRadians</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">toRadians-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">to_date</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">to_date-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">to_json</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_collection_functions.html">to_json-method</a></td>
+<td>Collection functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">to_timestamp</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">to_timestamp-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">to_utc_timestamp</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_diff_functions.html">to_utc_timestamp-method</a></td>
+<td>Date time arithmetic functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="mutate.html">transform</a></td>
+<td>Mutate</td></tr>
+<tr><td style="width: 25%;"><a href="mutate.html">transform-method</a></td>
+<td>Mutate</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">translate</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">translate-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">trim</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">trim-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">trunc</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">trunc-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+</table>
+
+<h2><a name="U">-- U --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_string_functions.html">unbase64</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">unbase64-method</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="uncacheTable.html">uncacheTable</a></td>
+<td>Uncache Table</td></tr>
+<tr><td style="width: 25%;"><a href="uncacheTable.html">uncacheTable.default</a></td>
+<td>Uncache Table</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">unhex</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_math_functions.html">unhex-method</a></td>
+<td>Math functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="union.html">union</a></td>
+<td>Return a new SparkDataFrame containing the union of rows</td></tr>
+<tr><td style="width: 25%;"><a href="union.html">union-method</a></td>
+<td>Return a new SparkDataFrame containing the union of rows</td></tr>
+<tr><td style="width: 25%;"><a href="union.html">unionAll</a></td>
+<td>Return a new SparkDataFrame containing the union of rows</td></tr>
+<tr><td style="width: 25%;"><a href="union.html">unionAll-method</a></td>
+<td>Return a new SparkDataFrame containing the union of rows</td></tr>
+<tr><td style="width: 25%;"><a href="unionByName.html">unionByName</a></td>
+<td>Return a new SparkDataFrame containing the union of rows, matched by column names</td></tr>
+<tr><td style="width: 25%;"><a href="unionByName.html">unionByName-method</a></td>
+<td>Return a new SparkDataFrame containing the union of rows, matched by column names</td></tr>
+<tr><td style="width: 25%;"><a href="distinct.html">unique</a></td>
+<td>Distinct</td></tr>
+<tr><td style="width: 25%;"><a href="distinct.html">unique-method</a></td>
+<td>Distinct</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">unix_timestamp</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">unix_timestamp-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="unpersist.html">unpersist</a></td>
+<td>Unpersist</td></tr>
+<tr><td style="width: 25%;"><a href="unpersist.html">unpersist-method</a></td>
+<td>Unpersist</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">upper</a></td>
+<td>String functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_string_functions.html">upper-method</a></td>
+<td>String functions for Column operations</td></tr>
+</table>
+
+<h2><a name="V">-- V --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">var</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">var-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">variance</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">variance-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">var_pop</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">var_pop-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">var_samp</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_aggregate_functions.html">var_samp-method</a></td>
+<td>Aggregate functions for Column operations</td></tr>
+</table>
+
+<h2><a name="W">-- W --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">weekofyear</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">weekofyear-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">when</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_nonaggregate_functions.html">when-method</a></td>
+<td>Non-aggregate functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="filter.html">where</a></td>
+<td>Filter</td></tr>
+<tr><td style="width: 25%;"><a href="filter.html">where-method</a></td>
+<td>Filter</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">window</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">window-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="windowOrderBy.html">windowOrderBy</a></td>
+<td>windowOrderBy</td></tr>
+<tr><td style="width: 25%;"><a href="windowOrderBy.html">windowOrderBy-method</a></td>
+<td>windowOrderBy</td></tr>
+<tr><td style="width: 25%;"><a href="windowPartitionBy.html">windowPartitionBy</a></td>
+<td>windowPartitionBy</td></tr>
+<tr><td style="width: 25%;"><a href="windowPartitionBy.html">windowPartitionBy-method</a></td>
+<td>windowPartitionBy</td></tr>
+<tr><td style="width: 25%;"><a href="WindowSpec.html">WindowSpec-class</a></td>
+<td>S4 class that represents a WindowSpec</td></tr>
+<tr><td style="width: 25%;"><a href="with.html">with</a></td>
+<td>Evaluate a R expression in an environment constructed from a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="with.html">with-method</a></td>
+<td>Evaluate a R expression in an environment constructed from a SparkDataFrame</td></tr>
+<tr><td style="width: 25%;"><a href="withColumn.html">withColumn</a></td>
+<td>WithColumn</td></tr>
+<tr><td style="width: 25%;"><a href="withColumn.html">withColumn-method</a></td>
+<td>WithColumn</td></tr>
+<tr><td style="width: 25%;"><a href="rename.html">withColumnRenamed</a></td>
+<td>rename</td></tr>
+<tr><td style="width: 25%;"><a href="rename.html">withColumnRenamed-method</a></td>
+<td>rename</td></tr>
+<tr><td style="width: 25%;"><a href="withWatermark.html">withWatermark</a></td>
+<td>withWatermark</td></tr>
+<tr><td style="width: 25%;"><a href="withWatermark.html">withWatermark-method</a></td>
+<td>withWatermark</td></tr>
+<tr><td style="width: 25%;"><a href="write.df.html">write.df</a></td>
+<td>Save the contents of SparkDataFrame to a data source.</td></tr>
+<tr><td style="width: 25%;"><a href="write.df.html">write.df-method</a></td>
+<td>Save the contents of SparkDataFrame to a data source.</td></tr>
+<tr><td style="width: 25%;"><a href="write.jdbc.html">write.jdbc</a></td>
+<td>Save the content of SparkDataFrame to an external database table via JDBC.</td></tr>
+<tr><td style="width: 25%;"><a href="write.jdbc.html">write.jdbc-method</a></td>
+<td>Save the content of SparkDataFrame to an external database table via JDBC.</td></tr>
+<tr><td style="width: 25%;"><a href="write.json.html">write.json</a></td>
+<td>Save the contents of SparkDataFrame as a JSON file</td></tr>
+<tr><td style="width: 25%;"><a href="write.json.html">write.json-method</a></td>
+<td>Save the contents of SparkDataFrame as a JSON file</td></tr>
+<tr><td style="width: 25%;"><a href="write.ml.html">write.ml</a></td>
+<td>Saves the MLlib model to the input path</td></tr>
+<tr><td style="width: 25%;"><a href="spark.als.html">write.ml-method</a></td>
+<td>Alternating Least Squares (ALS) for Collaborative Filtering</td></tr>
+<tr><td style="width: 25%;"><a href="spark.bisectingKmeans.html">write.ml-method</a></td>
+<td>Bisecting K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.decisionTree.html">write.ml-method</a></td>
+<td>Decision Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.fpGrowth.html">write.ml-method</a></td>
+<td>FP-growth</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gaussianMixture.html">write.ml-method</a></td>
+<td>Multivariate Gaussian Mixture Model (GMM)</td></tr>
+<tr><td style="width: 25%;"><a href="spark.gbt.html">write.ml-method</a></td>
+<td>Gradient Boosted Tree Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.glm.html">write.ml-method</a></td>
+<td>Generalized Linear Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.isoreg.html">write.ml-method</a></td>
+<td>Isotonic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.kmeans.html">write.ml-method</a></td>
+<td>K-Means Clustering Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.lda.html">write.ml-method</a></td>
+<td>Latent Dirichlet Allocation</td></tr>
+<tr><td style="width: 25%;"><a href="spark.logit.html">write.ml-method</a></td>
+<td>Logistic Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.mlp.html">write.ml-method</a></td>
+<td>Multilayer Perceptron Classification Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.naiveBayes.html">write.ml-method</a></td>
+<td>Naive Bayes Models</td></tr>
+<tr><td style="width: 25%;"><a href="spark.randomForest.html">write.ml-method</a></td>
+<td>Random Forest Model for Regression and Classification</td></tr>
+<tr><td style="width: 25%;"><a href="spark.survreg.html">write.ml-method</a></td>
+<td>Accelerated Failure Time (AFT) Survival Regression Model</td></tr>
+<tr><td style="width: 25%;"><a href="spark.svmLinear.html">write.ml-method</a></td>
+<td>Linear SVM Model</td></tr>
+<tr><td style="width: 25%;"><a href="write.orc.html">write.orc</a></td>
+<td>Save the contents of SparkDataFrame as an ORC file, preserving the schema.</td></tr>
+<tr><td style="width: 25%;"><a href="write.orc.html">write.orc-method</a></td>
+<td>Save the contents of SparkDataFrame as an ORC file, preserving the schema.</td></tr>
+<tr><td style="width: 25%;"><a href="write.parquet.html">write.parquet</a></td>
+<td>Save the contents of SparkDataFrame as a Parquet file, preserving the schema.</td></tr>
+<tr><td style="width: 25%;"><a href="write.parquet.html">write.parquet-method</a></td>
+<td>Save the contents of SparkDataFrame as a Parquet file, preserving the schema.</td></tr>
+<tr><td style="width: 25%;"><a href="write.stream.html">write.stream</a></td>
+<td>Write the streaming SparkDataFrame to a data source.</td></tr>
+<tr><td style="width: 25%;"><a href="write.stream.html">write.stream-method</a></td>
+<td>Write the streaming SparkDataFrame to a data source.</td></tr>
+<tr><td style="width: 25%;"><a href="write.text.html">write.text</a></td>
+<td>Save the content of SparkDataFrame in a text file at the specified path.</td></tr>
+<tr><td style="width: 25%;"><a href="write.text.html">write.text-method</a></td>
+<td>Save the content of SparkDataFrame in a text file at the specified path.</td></tr>
+</table>
+
+<h2><a name="Y">-- Y --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">year</a></td>
+<td>Date time functions for Column operations</td></tr>
+<tr><td style="width: 25%;"><a href="column_datetime_functions.html">year-method</a></td>
+<td>Date time functions for Column operations</td></tr>
+</table>
+
+<h2><a name="misc">-- misc --</a></h2>
+
+<table width="100%">
+<tr><td style="width: 25%;"><a href="not.html">!-method</a></td>
+<td>!</td></tr>
+<tr><td style="width: 25%;"><a href="select.html">$</a></td>
+<td>Select</td></tr>
+<tr><td style="width: 25%;"><a href="select.html">$-method</a></td>
+<td>Select</td></tr>
+<tr><td style="width: 25%;"><a href="select.html">$&lt;-</a></td>
+<td>Select</td></tr>
+<tr><td style="width: 25%;"><a href="select.html">$&lt;--method</a></td>
+<td>Select</td></tr>
+<tr><td style="width: 25%;"><a href="eq_null_safe.html">%&lt;=&gt;%</a></td>
+<td>%&lt;=&gt;%</td></tr>
+<tr><td style="width: 25%;"><a href="eq_null_safe.html">%&lt;=&gt;%-method</a></td>
+<td>%&lt;=&gt;%</td></tr>
+<tr><td style="width: 25%;"><a href="match.html">%in%</a></td>
+<td>Match a column with given values.</td></tr>
+<tr><td style="width: 25%;"><a href="match.html">%in%-method</a></td>
+<td>Match a column with given values.</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">[</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">[-method</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">[[</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">[[-method</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">[[&lt;-</a></td>
+<td>Subset</td></tr>
+<tr><td style="width: 25%;"><a href="subset.html">[[&lt;--method</a></td>
+<td>Subset</td></tr>
+</table>
 </body></html>
diff --git a/site/docs/2.4.3/api/R/00frame_toc.html b/site/docs/2.4.3/api/R/00frame_toc.html
index a8be285..4b40dfe 100644
--- a/site/docs/2.4.3/api/R/00frame_toc.html
+++ b/site/docs/2.4.3/api/R/00frame_toc.html
@@ -86,6 +86,239 @@ dt, p code {
 <body>
 <h1>SparkR</h1>
 
+<ul>
+<li><a href="AFTSurvivalRegressionModel-class.html" target="content">AFTSurvivalRegressionModel-class</a></li>
+<li><a href="ALSModel-class.html" target="content">ALSModel-class</a></li>
+<li><a href="BisectingKMeansModel-class.html" target="content">BisectingKMeansModel-class</a></li>
+<li><a href="DecisionTreeClassificationModel-class.html" target="content">DecisionTreeClassificationModel-class</a></li>
+<li><a href="DecisionTreeRegressionModel-class.html" target="content">DecisionTreeRegressionModel-class</a></li>
+<li><a href="FPGrowthModel-class.html" target="content">FPGrowthModel-class</a></li>
+<li><a href="GBTClassificationModel-class.html" target="content">GBTClassificationModel-class</a></li>
+<li><a href="GBTRegressionModel-class.html" target="content">GBTRegressionModel-class</a></li>
+<li><a href="GaussianMixtureModel-class.html" target="content">GaussianMixtureModel-class</a></li>
+<li><a href="GeneralizedLinearRegressionModel-class.html" target="content">GeneralizedLinearRegressionModel-class</a></li>
+<li><a href="GroupedData.html" target="content">GroupedData</a></li>
+<li><a href="IsotonicRegressionModel-class.html" target="content">IsotonicRegressionModel-class</a></li>
+<li><a href="KMeansModel-class.html" target="content">KMeansModel-class</a></li>
+<li><a href="KSTest-class.html" target="content">KSTest-class</a></li>
+<li><a href="LDAModel-class.html" target="content">LDAModel-class</a></li>
+<li><a href="LinearSVCModel-class.html" target="content">LinearSVCModel-class</a></li>
+<li><a href="LogisticRegressionModel-class.html" target="content">LogisticRegressionModel-class</a></li>
+<li><a href="MultilayerPerceptronClassificationModel-class.html" target="content">MultilayerPerceptronClassificationModel-class</a></li>
+<li><a href="NaiveBayesModel-class.html" target="content">NaiveBayesModel-class</a></li>
+<li><a href="RandomForestClassificationModel-class.html" target="content">RandomForestClassificationModel-class</a></li>
+<li><a href="RandomForestRegressionModel-class.html" target="content">RandomForestRegressionModel-class</a></li>
+<li><a href="SparkDataFrame.html" target="content">SparkDataFrame</a></li>
+<li><a href="StreamingQuery.html" target="content">StreamingQuery</a></li>
+<li><a href="WindowSpec.html" target="content">WindowSpec</a></li>
+<li><a href="alias.html" target="content">alias</a></li>
+<li><a href="approxQuantile.html" target="content">approxQuantile</a></li>
+<li><a href="arrange.html" target="content">arrange</a></li>
+<li><a href="as.data.frame.html" target="content">as.data.frame</a></li>
+<li><a href="attach.html" target="content">attach</a></li>
+<li><a href="avg.html" target="content">avg</a></li>
+<li><a href="awaitTermination.html" target="content">awaitTermination</a></li>
+<li><a href="between.html" target="content">between</a></li>
+<li><a href="broadcast.html" target="content">broadcast</a></li>
+<li><a href="cache.html" target="content">cache</a></li>
+<li><a href="cacheTable.html" target="content">cacheTable</a></li>
+<li><a href="cancelJobGroup.html" target="content">cancelJobGroup</a></li>
+<li><a href="cast.html" target="content">cast</a></li>
+<li><a href="checkpoint.html" target="content">checkpoint</a></li>
+<li><a href="clearCache.html" target="content">clearCache</a></li>
+<li><a href="clearJobGroup.html" target="content">clearJobGroup</a></li>
+<li><a href="coalesce.html" target="content">coalesce</a></li>
+<li><a href="collect.html" target="content">collect</a></li>
+<li><a href="coltypes.html" target="content">coltypes</a></li>
+<li><a href="column.html" target="content">column</a></li>
+<li><a href="column_aggregate_functions.html" target="content">column<em>aggregate</em>functions</a></li>
+<li><a href="column_collection_functions.html" target="content">column<em>collection</em>functions</a></li>
+<li><a href="column_datetime_diff_functions.html" target="content">column<em>datetime</em>diff_functions</a></li>
+<li><a href="column_datetime_functions.html" target="content">column<em>datetime</em>functions</a></li>
+<li><a href="column_math_functions.html" target="content">column<em>math</em>functions</a></li>
+<li><a href="column_misc_functions.html" target="content">column<em>misc</em>functions</a></li>
+<li><a href="column_nonaggregate_functions.html" target="content">column<em>nonaggregate</em>functions</a></li>
+<li><a href="column_string_functions.html" target="content">column<em>string</em>functions</a></li>
+<li><a href="column_window_functions.html" target="content">column<em>window</em>functions</a></li>
+<li><a href="columnfunctions.html" target="content">columnfunctions</a></li>
+<li><a href="columns.html" target="content">columns</a></li>
+<li><a href="corr.html" target="content">corr</a></li>
+<li><a href="count.html" target="content">count</a></li>
+<li><a href="cov.html" target="content">cov</a></li>
+<li><a href="createDataFrame.html" target="content">createDataFrame</a></li>
+<li><a href="createExternalTable-deprecated.html" target="content">createExternalTable-deprecated</a></li>
+<li><a href="createOrReplaceTempView.html" target="content">createOrReplaceTempView</a></li>
+<li><a href="createTable.html" target="content">createTable</a></li>
+<li><a href="crossJoin.html" target="content">crossJoin</a></li>
+<li><a href="crosstab.html" target="content">crosstab</a></li>
+<li><a href="cube.html" target="content">cube</a></li>
+<li><a href="currentDatabase.html" target="content">currentDatabase</a></li>
+<li><a href="dapply.html" target="content">dapply</a></li>
+<li><a href="dapplyCollect.html" target="content">dapplyCollect</a></li>
+<li><a href="describe.html" target="content">describe</a></li>
+<li><a href="dim.html" target="content">dim</a></li>
+<li><a href="distinct.html" target="content">distinct</a></li>
+<li><a href="drop.html" target="content">drop</a></li>
+<li><a href="dropDuplicates.html" target="content">dropDuplicates</a></li>
+<li><a href="dropTempTable-deprecated.html" target="content">dropTempTable-deprecated</a></li>
+<li><a href="dropTempView.html" target="content">dropTempView</a></li>
+<li><a href="dtypes.html" target="content">dtypes</a></li>
+<li><a href="endsWith.html" target="content">endsWith</a></li>
+<li><a href="eq_null_safe.html" target="content">eq<em>null</em>safe</a></li>
+<li><a href="except.html" target="content">except</a></li>
+<li><a href="exceptAll.html" target="content">exceptAll</a></li>
+<li><a href="explain.html" target="content">explain</a></li>
+<li><a href="filter.html" target="content">filter</a></li>
+<li><a href="first.html" target="content">first</a></li>
+<li><a href="fitted.html" target="content">fitted</a></li>
+<li><a href="freqItems.html" target="content">freqItems</a></li>
+<li><a href="gapply.html" target="content">gapply</a></li>
+<li><a href="gapplyCollect.html" target="content">gapplyCollect</a></li>
+<li><a href="getLocalProperty.html" target="content">getLocalProperty</a></li>
+<li><a href="getNumPartitions.html" target="content">getNumPartitions</a></li>
+<li><a href="glm.html" target="content">glm</a></li>
+<li><a href="groupBy.html" target="content">groupBy</a></li>
+<li><a href="hashCode.html" target="content">hashCode</a></li>
+<li><a href="head.html" target="content">head</a></li>
+<li><a href="hint.html" target="content">hint</a></li>
+<li><a href="histogram.html" target="content">histogram</a></li>
+<li><a href="insertInto.html" target="content">insertInto</a></li>
+<li><a href="install.spark.html" target="content">install.spark</a></li>
+<li><a href="intersect.html" target="content">intersect</a></li>
+<li><a href="intersectAll.html" target="content">intersectAll</a></li>
+<li><a href="isActive.html" target="content">isActive</a></li>
+<li><a href="isLocal.html" target="content">isLocal</a></li>
+<li><a href="isStreaming.html" target="content">isStreaming</a></li>
+<li><a href="join.html" target="content">join</a></li>
+<li><a href="last.html" target="content">last</a></li>
+<li><a href="lastProgress.html" target="content">lastProgress</a></li>
+<li><a href="limit.html" target="content">limit</a></li>
+<li><a href="listColumns.html" target="content">listColumns</a></li>
+<li><a href="listDatabases.html" target="content">listDatabases</a></li>
+<li><a href="listFunctions.html" target="content">listFunctions</a></li>
+<li><a href="listTables.html" target="content">listTables</a></li>
+<li><a href="localCheckpoint.html" target="content">localCheckpoint</a></li>
+<li><a href="match.html" target="content">match</a></li>
+<li><a href="merge.html" target="content">merge</a></li>
+<li><a href="mutate.html" target="content">mutate</a></li>
+<li><a href="nafunctions.html" target="content">nafunctions</a></li>
+<li><a href="ncol.html" target="content">ncol</a></li>
+<li><a href="not.html" target="content">not</a></li>
+<li><a href="nrow.html" target="content">nrow</a></li>
+<li><a href="orderBy.html" target="content">orderBy</a></li>
+<li><a href="otherwise.html" target="content">otherwise</a></li>
+<li><a href="over.html" target="content">over</a></li>
+<li><a href="partitionBy.html" target="content">partitionBy</a></li>
+<li><a href="persist.html" target="content">persist</a></li>
+<li><a href="pivot.html" target="content">pivot</a></li>
+<li><a href="predict.html" target="content">predict</a></li>
+<li><a href="print.jobj.html" target="content">print.jobj</a></li>
+<li><a href="print.structField.html" target="content">print.structField</a></li>
+<li><a href="print.structType.html" target="content">print.structType</a></li>
+<li><a href="printSchema.html" target="content">printSchema</a></li>
+<li><a href="queryName.html" target="content">queryName</a></li>
+<li><a href="randomSplit.html" target="content">randomSplit</a></li>
+<li><a href="rangeBetween.html" target="content">rangeBetween</a></li>
+<li><a href="rbind.html" target="content">rbind</a></li>
+<li><a href="read.df.html" target="content">read.df</a></li>
+<li><a href="read.jdbc.html" target="content">read.jdbc</a></li>
+<li><a href="read.json.html" target="content">read.json</a></li>
+<li><a href="read.ml.html" target="content">read.ml</a></li>
+<li><a href="read.orc.html" target="content">read.orc</a></li>
+<li><a href="read.parquet.html" target="content">read.parquet</a></li>
+<li><a href="read.stream.html" target="content">read.stream</a></li>
+<li><a href="read.text.html" target="content">read.text</a></li>
+<li><a href="recoverPartitions.html" target="content">recoverPartitions</a></li>
+<li><a href="refreshByPath.html" target="content">refreshByPath</a></li>
+<li><a href="refreshTable.html" target="content">refreshTable</a></li>
+<li><a href="registerTempTable-deprecated.html" target="content">registerTempTable-deprecated</a></li>
+<li><a href="rename.html" target="content">rename</a></li>
+<li><a href="repartition.html" target="content">repartition</a></li>
+<li><a href="repartitionByRange.html" target="content">repartitionByRange</a></li>
+<li><a href="rollup.html" target="content">rollup</a></li>
+<li><a href="rowsBetween.html" target="content">rowsBetween</a></li>
+<li><a href="sample.html" target="content">sample</a></li>
+<li><a href="sampleBy.html" target="content">sampleBy</a></li>
+<li><a href="saveAsTable.html" target="content">saveAsTable</a></li>
+<li><a href="schema.html" target="content">schema</a></li>
+<li><a href="select.html" target="content">select</a></li>
+<li><a href="selectExpr.html" target="content">selectExpr</a></li>
+<li><a href="setCheckpointDir.html" target="content">setCheckpointDir</a></li>
+<li><a href="setCurrentDatabase.html" target="content">setCurrentDatabase</a></li>
+<li><a href="setJobDescription.html" target="content">setJobDescription</a></li>
+<li><a href="setJobGroup.html" target="content">setJobGroup</a></li>
+<li><a href="setLocalProperty.html" target="content">setLocalProperty</a></li>
+<li><a href="setLogLevel.html" target="content">setLogLevel</a></li>
+<li><a href="show.html" target="content">show</a></li>
+<li><a href="showDF.html" target="content">showDF</a></li>
+<li><a href="spark.addFile.html" target="content">spark.addFile</a></li>
+<li><a href="spark.als.html" target="content">spark.als</a></li>
+<li><a href="spark.bisectingKmeans.html" target="content">spark.bisectingKmeans</a></li>
+<li><a href="spark.decisionTree.html" target="content">spark.decisionTree</a></li>
+<li><a href="spark.fpGrowth.html" target="content">spark.fpGrowth</a></li>
+<li><a href="spark.gaussianMixture.html" target="content">spark.gaussianMixture</a></li>
+<li><a href="spark.gbt.html" target="content">spark.gbt</a></li>
+<li><a href="spark.getSparkFiles.html" target="content">spark.getSparkFiles</a></li>
+<li><a href="spark.getSparkFilesRootDirectory.html" target="content">spark.getSparkFilesRootDirectory</a></li>
+<li><a href="spark.glm.html" target="content">spark.glm</a></li>
+<li><a href="spark.isoreg.html" target="content">spark.isoreg</a></li>
+<li><a href="spark.kmeans.html" target="content">spark.kmeans</a></li>
+<li><a href="spark.kstest.html" target="content">spark.kstest</a></li>
+<li><a href="spark.lapply.html" target="content">spark.lapply</a></li>
+<li><a href="spark.lda.html" target="content">spark.lda</a></li>
+<li><a href="spark.logit.html" target="content">spark.logit</a></li>
+<li><a href="spark.mlp.html" target="content">spark.mlp</a></li>
+<li><a href="spark.naiveBayes.html" target="content">spark.naiveBayes</a></li>
+<li><a href="spark.randomForest.html" target="content">spark.randomForest</a></li>
+<li><a href="spark.survreg.html" target="content">spark.survreg</a></li>
+<li><a href="spark.svmLinear.html" target="content">spark.svmLinear</a></li>
+<li><a href="sparkR.callJMethod.html" target="content">sparkR.callJMethod</a></li>
+<li><a href="sparkR.callJStatic.html" target="content">sparkR.callJStatic</a></li>
+<li><a href="sparkR.conf.html" target="content">sparkR.conf</a></li>
+<li><a href="sparkR.init-deprecated.html" target="content">sparkR.init-deprecated</a></li>
+<li><a href="sparkR.newJObject.html" target="content">sparkR.newJObject</a></li>
+<li><a href="sparkR.session.html" target="content">sparkR.session</a></li>
+<li><a href="sparkR.session.stop.html" target="content">sparkR.session.stop</a></li>
+<li><a href="sparkR.uiWebUrl.html" target="content">sparkR.uiWebUrl</a></li>
+<li><a href="sparkR.version.html" target="content">sparkR.version</a></li>
+<li><a href="sparkRHive.init-deprecated.html" target="content">sparkRHive.init-deprecated</a></li>
+<li><a href="sparkRSQL.init-deprecated.html" target="content">sparkRSQL.init-deprecated</a></li>
+<li><a href="sql.html" target="content">sql</a></li>
+<li><a href="startsWith.html" target="content">startsWith</a></li>
+<li><a href="status.html" target="content">status</a></li>
+<li><a href="stopQuery.html" target="content">stopQuery</a></li>
+<li><a href="storageLevel.html" target="content">storageLevel</a></li>
+<li><a href="str.html" target="content">str</a></li>
+<li><a href="structField.html" target="content">structField</a></li>
+<li><a href="structType.html" target="content">structType</a></li>
+<li><a href="subset.html" target="content">subset</a></li>
+<li><a href="substr.html" target="content">substr</a></li>
+<li><a href="summarize.html" target="content">summarize</a></li>
+<li><a href="summary.html" target="content">summary</a></li>
+<li><a href="tableNames.html" target="content">tableNames</a></li>
+<li><a href="tableToDF.html" target="content">tableToDF</a></li>
+<li><a href="tables.html" target="content">tables</a></li>
+<li><a href="take.html" target="content">take</a></li>
+<li><a href="toJSON.html" target="content">toJSON</a></li>
+<li><a href="uncacheTable.html" target="content">uncacheTable</a></li>
+<li><a href="union.html" target="content">union</a></li>
+<li><a href="unionByName.html" target="content">unionByName</a></li>
+<li><a href="unpersist.html" target="content">unpersist</a></li>
+<li><a href="windowOrderBy.html" target="content">windowOrderBy</a></li>
+<li><a href="windowPartitionBy.html" target="content">windowPartitionBy</a></li>
+<li><a href="with.html" target="content">with</a></li>
+<li><a href="withColumn.html" target="content">withColumn</a></li>
+<li><a href="withWatermark.html" target="content">withWatermark</a></li>
+<li><a href="write.df.html" target="content">write.df</a></li>
+<li><a href="write.jdbc.html" target="content">write.jdbc</a></li>
+<li><a href="write.json.html" target="content">write.json</a></li>
+<li><a href="write.ml.html" target="content">write.ml</a></li>
+<li><a href="write.orc.html" target="content">write.orc</a></li>
+<li><a href="write.parquet.html" target="content">write.parquet</a></li>
+<li><a href="write.stream.html" target="content">write.stream</a></li>
+<li><a href="write.text.html" target="content">write.text</a></li>
+</ul>
+
 <p>Generated with <a href="https://yihui.name/knitr">knitr</a>  1.22</p>
 
 </body>
diff --git a/site/docs/2.4.3/api/R/AFTSurvivalRegressionModel-class.html b/site/docs/2.4.3/api/R/AFTSurvivalRegressionModel-class.html
new file mode 100644
index 0000000..2a91180
--- /dev/null
+++ b/site/docs/2.4.3/api/R/AFTSurvivalRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a AFTSurvivalRegressionModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for AFTSurvivalRegressionModel-class {SparkR}"><tr><td>AFTSurvivalRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a AFTSurvivalRegressionModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a AFTSurvivalRegressionModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala AFTSurvivalRegressionWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>AFTSurvivalRegressionModel since 2.0.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/ALSModel-class.html b/site/docs/2.4.3/api/R/ALSModel-class.html
new file mode 100644
index 0000000..3ee9ea6
--- /dev/null
+++ b/site/docs/2.4.3/api/R/ALSModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents an ALSModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for ALSModel-class {SparkR}"><tr><td>ALSModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents an ALSModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents an ALSModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala ALSWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>ALSModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/BisectingKMeansModel-class.html b/site/docs/2.4.3/api/R/BisectingKMeansModel-class.html
new file mode 100644
index 0000000..de08f86
--- /dev/null
+++ b/site/docs/2.4.3/api/R/BisectingKMeansModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a BisectingKMeansModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for BisectingKMeansModel-class {SparkR}"><tr><td>BisectingKMeansModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a BisectingKMeansModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a BisectingKMeansModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala BisectingKMeansModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>BisectingKMeansModel since 2.2.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/DecisionTreeClassificationModel-class.html b/site/docs/2.4.3/api/R/DecisionTreeClassificationModel-class.html
new file mode 100644
index 0000000..a315d7f
--- /dev/null
+++ b/site/docs/2.4.3/api/R/DecisionTreeClassificationModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a DecisionTreeClassificationModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for DecisionTreeClassificationModel-class {SparkR}"><tr><td>DecisionTreeClassificationModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a DecisionTreeClassificationModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a DecisionTreeClassificationModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala DecisionTreeClassificationModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>DecisionTreeClassificationModel since 2.3.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/DecisionTreeRegressionModel-class.html b/site/docs/2.4.3/api/R/DecisionTreeRegressionModel-class.html
new file mode 100644
index 0000000..7f4b279
--- /dev/null
+++ b/site/docs/2.4.3/api/R/DecisionTreeRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a DecisionTreeRegressionModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for DecisionTreeRegressionModel-class {SparkR}"><tr><td>DecisionTreeRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a DecisionTreeRegressionModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a DecisionTreeRegressionModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala DecisionTreeRegressionModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>DecisionTreeRegressionModel since 2.3.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/FPGrowthModel-class.html b/site/docs/2.4.3/api/R/FPGrowthModel-class.html
new file mode 100644
index 0000000..01b3b92
--- /dev/null
+++ b/site/docs/2.4.3/api/R/FPGrowthModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a FPGrowthModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for FPGrowthModel-class {SparkR}"><tr><td>FPGrowthModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a FPGrowthModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a FPGrowthModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala FPGrowthModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>FPGrowthModel since 2.2.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/GBTClassificationModel-class.html b/site/docs/2.4.3/api/R/GBTClassificationModel-class.html
new file mode 100644
index 0000000..e0efd69
--- /dev/null
+++ b/site/docs/2.4.3/api/R/GBTClassificationModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a GBTClassificationModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for GBTClassificationModel-class {SparkR}"><tr><td>GBTClassificationModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a GBTClassificationModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a GBTClassificationModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala GBTClassificationModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>GBTClassificationModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/GBTRegressionModel-class.html b/site/docs/2.4.3/api/R/GBTRegressionModel-class.html
new file mode 100644
index 0000000..ab824c3
--- /dev/null
+++ b/site/docs/2.4.3/api/R/GBTRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a GBTRegressionModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for GBTRegressionModel-class {SparkR}"><tr><td>GBTRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a GBTRegressionModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a GBTRegressionModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala GBTRegressionModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>GBTRegressionModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/GaussianMixtureModel-class.html b/site/docs/2.4.3/api/R/GaussianMixtureModel-class.html
new file mode 100644
index 0000000..88073cb
--- /dev/null
+++ b/site/docs/2.4.3/api/R/GaussianMixtureModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a GaussianMixtureModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for GaussianMixtureModel-class {SparkR}"><tr><td>GaussianMixtureModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a GaussianMixtureModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a GaussianMixtureModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala GaussianMixtureModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>GaussianMixtureModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/GeneralizedLinearRegressionModel-class.html b/site/docs/2.4.3/api/R/GeneralizedLinearRegressionModel-class.html
new file mode 100644
index 0000000..dbb3609
--- /dev/null
+++ b/site/docs/2.4.3/api/R/GeneralizedLinearRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a generalized linear model</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for GeneralizedLinearRegressionModel-class {SparkR}"><tr><td>GeneralizedLinearRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a generalized linear model</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a generalized linear model
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala GeneralizedLinearRegressionWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>GeneralizedLinearRegressionModel since 2.0.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/GroupedData.html b/site/docs/2.4.3/api/R/GroupedData.html
new file mode 100644
index 0000000..73d8f6f
--- /dev/null
+++ b/site/docs/2.4.3/api/R/GroupedData.html
@@ -0,0 +1,45 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a GroupedData</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for GroupedData-class {SparkR}"><tr><td>GroupedData-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a GroupedData</h2>
+
+<h3>Description</h3>
+
+<p>GroupedDatas can be created using groupBy() on a SparkDataFrame
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+groupedData(sgd)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>sgd</code></td>
+<td>
+<p>A Java object reference to the backing Scala GroupedData</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>GroupedData since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>groupBy
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/IsotonicRegressionModel-class.html b/site/docs/2.4.3/api/R/IsotonicRegressionModel-class.html
new file mode 100644
index 0000000..84f4240
--- /dev/null
+++ b/site/docs/2.4.3/api/R/IsotonicRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents an IsotonicRegressionModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for IsotonicRegressionModel-class {SparkR}"><tr><td>IsotonicRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents an IsotonicRegressionModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents an IsotonicRegressionModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala IsotonicRegressionModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>IsotonicRegressionModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/KMeansModel-class.html b/site/docs/2.4.3/api/R/KMeansModel-class.html
new file mode 100644
index 0000000..9656597
--- /dev/null
+++ b/site/docs/2.4.3/api/R/KMeansModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a KMeansModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for KMeansModel-class {SparkR}"><tr><td>KMeansModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a KMeansModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a KMeansModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala KMeansModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>KMeansModel since 2.0.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/KSTest-class.html b/site/docs/2.4.3/api/R/KSTest-class.html
new file mode 100644
index 0000000..d2fdcaa
--- /dev/null
+++ b/site/docs/2.4.3/api/R/KSTest-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents an KSTest</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for KSTest-class {SparkR}"><tr><td>KSTest-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents an KSTest</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents an KSTest
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala KSTestWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>KSTest since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/LDAModel-class.html b/site/docs/2.4.3/api/R/LDAModel-class.html
new file mode 100644
index 0000000..de95170
--- /dev/null
+++ b/site/docs/2.4.3/api/R/LDAModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents an LDAModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for LDAModel-class {SparkR}"><tr><td>LDAModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents an LDAModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents an LDAModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala LDAWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>LDAModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/LinearSVCModel-class.html b/site/docs/2.4.3/api/R/LinearSVCModel-class.html
new file mode 100644
index 0000000..76a3dbf
--- /dev/null
+++ b/site/docs/2.4.3/api/R/LinearSVCModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents an LinearSVCModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for LinearSVCModel-class {SparkR}"><tr><td>LinearSVCModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents an LinearSVCModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents an LinearSVCModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala LinearSVCModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>LinearSVCModel since 2.2.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/LogisticRegressionModel-class.html b/site/docs/2.4.3/api/R/LogisticRegressionModel-class.html
new file mode 100644
index 0000000..b7e0d74
--- /dev/null
+++ b/site/docs/2.4.3/api/R/LogisticRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents an LogisticRegressionModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for LogisticRegressionModel-class {SparkR}"><tr><td>LogisticRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents an LogisticRegressionModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents an LogisticRegressionModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala LogisticRegressionModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>LogisticRegressionModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/MultilayerPerceptronClassificationModel-class.html b/site/docs/2.4.3/api/R/MultilayerPerceptronClassificationModel-class.html
new file mode 100644
index 0000000..a626671
--- /dev/null
+++ b/site/docs/2.4.3/api/R/MultilayerPerceptronClassificationModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for MultilayerPerceptronClassificationModel-class {SparkR}"><tr><td>MultilayerPerceptronClassificationModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a MultilayerPerceptronClassificationModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a MultilayerPerceptronClassificationModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala MultilayerPerceptronClassifierWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>MultilayerPerceptronClassificationModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/NaiveBayesModel-class.html b/site/docs/2.4.3/api/R/NaiveBayesModel-class.html
new file mode 100644
index 0000000..3b874e5
--- /dev/null
+++ b/site/docs/2.4.3/api/R/NaiveBayesModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a NaiveBayesModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for NaiveBayesModel-class {SparkR}"><tr><td>NaiveBayesModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a NaiveBayesModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a NaiveBayesModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala NaiveBayesWrapper</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>NaiveBayesModel since 2.0.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/RandomForestClassificationModel-class.html b/site/docs/2.4.3/api/R/RandomForestClassificationModel-class.html
new file mode 100644
index 0000000..480017b
--- /dev/null
+++ b/site/docs/2.4.3/api/R/RandomForestClassificationModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a RandomForestClassificationModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for RandomForestClassificationModel-class {SparkR}"><tr><td>RandomForestClassificationModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a RandomForestClassificationModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a RandomForestClassificationModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala RandomForestClassificationModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>RandomForestClassificationModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/RandomForestRegressionModel-class.html b/site/docs/2.4.3/api/R/RandomForestRegressionModel-class.html
new file mode 100644
index 0000000..d2a5b59
--- /dev/null
+++ b/site/docs/2.4.3/api/R/RandomForestRegressionModel-class.html
@@ -0,0 +1,32 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a RandomForestRegressionModel</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for RandomForestRegressionModel-class {SparkR}"><tr><td>RandomForestRegressionModel-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a RandomForestRegressionModel</h2>
+
+<h3>Description</h3>
+
+<p>S4 class that represents a RandomForestRegressionModel
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>jobj</code></td>
+<td>
+<p>a Java object reference to the backing Scala RandomForestRegressionModel</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>RandomForestRegressionModel since 2.1.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/SparkDataFrame.html b/site/docs/2.4.3/api/R/SparkDataFrame.html
new file mode 100644
index 0000000..3fd4c71
--- /dev/null
+++ b/site/docs/2.4.3/api/R/SparkDataFrame.html
@@ -0,0 +1,102 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a SparkDataFrame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for SparkDataFrame-class {SparkR}"><tr><td>SparkDataFrame-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>SparkDataFrames can be created using functions like <a href="createDataFrame.html">createDataFrame</a>,
+<a href="read.json.html">read.json</a>, <a href="../../base/html/table.html">table</a> etc.
+</p>
+
+
+<h3>Slots</h3>
+
+
+<dl>
+<dt><code>env</code></dt><dd><p>An R environment that stores bookkeeping states of the SparkDataFrame</p>
+</dd>
+<dt><code>sdf</code></dt><dd><p>A Java object reference to the backing Scala DataFrame</p>
+</dd>
+</dl>
+
+
+<h3>Note</h3>
+
+<p>SparkDataFrame since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="createDataFrame.html">createDataFrame</a>, <a href="read.json.html">read.json</a>, <a href="../../base/html/table.html">table</a>
+</p>
+<p><a href="https://spark.apache.org/docs/latest/sparkr.html#sparkr-dataframes">https://spark.apache.org/docs/latest/sparkr.html#sparkr-dataframes</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="summarize.html">agg</a></code>,
+<code><a href="alias.html">alias</a></code>, <code><a href="arrange.html">arrange</a></code>,
+<code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df &lt;- createDataFrame(faithful)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/StreamingQuery.html b/site/docs/2.4.3/api/R/StreamingQuery.html
new file mode 100644
index 0000000..1d771c1
--- /dev/null
+++ b/site/docs/2.4.3/api/R/StreamingQuery.html
@@ -0,0 +1,40 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a StreamingQuery</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for StreamingQuery-class {SparkR}"><tr><td>StreamingQuery-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a StreamingQuery</h2>
+
+<h3>Description</h3>
+
+<p>StreamingQuery can be created by using read.stream() and write.stream()
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>ssq</code></td>
+<td>
+<p>A Java object reference to the backing Scala StreamingQuery</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>StreamingQuery since 2.2.0
+</p>
+<p>experimental
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="read.stream.html">read.stream</a>
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/WindowSpec.html b/site/docs/2.4.3/api/R/WindowSpec.html
new file mode 100644
index 0000000..dfc23e1
--- /dev/null
+++ b/site/docs/2.4.3/api/R/WindowSpec.html
@@ -0,0 +1,38 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a WindowSpec</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for WindowSpec-class {SparkR}"><tr><td>WindowSpec-class {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a WindowSpec</h2>
+
+<h3>Description</h3>
+
+<p>WindowSpec can be created by using windowPartitionBy() or windowOrderBy()
+</p>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>sws</code></td>
+<td>
+<p>A Java object reference to the backing Scala WindowSpec</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>WindowSpec since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="windowPartitionBy.html">windowPartitionBy</a>, <a href="windowOrderBy.html">windowOrderBy</a>
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/alias.html b/site/docs/2.4.3/api/R/alias.html
new file mode 100644
index 0000000..f368bd8
--- /dev/null
+++ b/site/docs/2.4.3/api/R/alias.html
@@ -0,0 +1,134 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: alias</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for alias {SparkR}"><tr><td>alias {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>alias</h2>
+
+<h3>Description</h3>
+
+<p>Returns a new SparkDataFrame or a Column with an alias set. Equivalent to SQL &quot;AS&quot; keyword.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'Column'
+alias(object, data)
+
+## S4 method for signature 'SparkDataFrame'
+alias(object, data)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>x a SparkDataFrame or a Column</p>
+</td></tr>
+<tr valign="top"><td><code>data</code></td>
+<td>
+<p>new name to use</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>a SparkDataFrame or a Column
+</p>
+
+
+<h3>Note</h3>
+
+<p>alias(Column) since 1.4.0
+</p>
+<p>alias(SparkDataFrame) since 2.3.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="between.html">between</a></code>,
+<code><a href="cast.html">cast</a></code>, <code><a href="endsWith.html">endsWith</a></code>,
+<code><a href="otherwise.html">otherwise</a></code>, <code><a href="over.html">over</a></code>,
+<code><a href="startsWith.html">startsWith</a></code>, <code><a href="substr.html">substr</a></code>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="arrange.html">arrange</a></code>,
+<code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(iris)
+##D 
+##D head(select(
+##D   df, alias(df$Sepal_Length, &quot;slength&quot;), alias(df$Petal_Length, &quot;plength&quot;)
+##D ))
+## End(Not run)
+## Not run: 
+##D df &lt;- alias(createDataFrame(mtcars), &quot;mtcars&quot;)
+##D avg_mpg &lt;- alias(agg(groupBy(df, df$cyl), avg(df$mpg)), &quot;avg_mpg&quot;)
+##D 
+##D head(select(df, column(&quot;mtcars.mpg&quot;)))
+##D head(join(df, avg_mpg, column(&quot;mtcars.cyl&quot;) == column(&quot;avg_mpg.cyl&quot;)))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/approxQuantile.html b/site/docs/2.4.3/api/R/approxQuantile.html
new file mode 100644
index 0000000..3a6bec9
--- /dev/null
+++ b/site/docs/2.4.3/api/R/approxQuantile.html
@@ -0,0 +1,98 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Calculates the approximate quantiles of numerical columns of...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for approxQuantile {SparkR}"><tr><td>approxQuantile {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Calculates the approximate quantiles of numerical columns of a SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>Calculates the approximate quantiles of numerical columns of a SparkDataFrame.
+The result of this algorithm has the following deterministic bound:
+If the SparkDataFrame has N elements and if we request the quantile at probability p up to
+error err, then the algorithm will return a sample x from the SparkDataFrame so that the
+*exact* rank of x is close to (p * N). More precisely,
+floor((p - err) * N) &lt;= rank(x) &lt;= ceil((p + err) * N).
+This method implements a variation of the Greenwald-Khanna algorithm (with some speed
+optimizations). The algorithm was first present in [[http://dx.doi.org/10.1145/375663.375670
+Space-efficient Online Computation of Quantile Summaries]] by Greenwald and Khanna.
+Note that NA values will be ignored in numerical columns before calculation. For
+columns only containing NA values, an empty list is returned.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'SparkDataFrame,character,numeric,numeric'
+approxQuantile(x,
+  cols, probabilities, relativeError)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>cols</code></td>
+<td>
+<p>A single column name, or a list of names for multiple columns.</p>
+</td></tr>
+<tr valign="top"><td><code>probabilities</code></td>
+<td>
+<p>A list of quantile probabilities. Each number must belong to [0, 1].
+For example 0 is the minimum, 0.5 is the median, 1 is the maximum.</p>
+</td></tr>
+<tr valign="top"><td><code>relativeError</code></td>
+<td>
+<p>The relative target precision to achieve (&gt;= 0). If set to zero,
+the exact quantiles are computed, which could be very expensive.
+Note that values greater than 1 are accepted but give the same result as 1.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>The approximate quantiles at the given probabilities. If the input is a single column
+name, the output is a list of approximate quantiles in that column; If the input is
+multiple column names, the output should be a list, and each element in it is a list of
+numeric values which represents the approximate quantiles in corresponding column.
+</p>
+
+
+<h3>Note</h3>
+
+<p>approxQuantile since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other stat functions: <code><a href="corr.html">corr</a></code>,
+<code><a href="cov.html">cov</a></code>, <code><a href="crosstab.html">crosstab</a></code>,
+<code><a href="freqItems.html">freqItems</a></code>, <code><a href="sampleBy.html">sampleBy</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- read.json(&quot;/path/to/file.json&quot;)
+##D quantiles &lt;- approxQuantile(df, &quot;key&quot;, c(0.5, 0.8), 0.0)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/arrange.html b/site/docs/2.4.3/api/R/arrange.html
new file mode 100644
index 0000000..fb1f6ef
--- /dev/null
+++ b/site/docs/2.4.3/api/R/arrange.html
@@ -0,0 +1,147 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Arrange Rows by Variables</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for arrange {SparkR}"><tr><td>arrange {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Arrange Rows by Variables</h2>
+
+<h3>Description</h3>
+
+<p>Sort a SparkDataFrame by the specified column(s).
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+arrange(x, col, ...)
+
+## S4 method for signature 'SparkDataFrame,Column'
+arrange(x, col, ...,
+  withinPartitions = FALSE)
+
+## S4 method for signature 'SparkDataFrame,character'
+arrange(x, col, ...,
+  decreasing = FALSE, withinPartitions = FALSE)
+
+## S4 method for signature 'SparkDataFrame,characterOrColumn'
+orderBy(x, col, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame to be sorted.</p>
+</td></tr>
+<tr valign="top"><td><code>col</code></td>
+<td>
+<p>a character or Column object indicating the fields to sort on</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional sorting fields</p>
+</td></tr>
+<tr valign="top"><td><code>withinPartitions</code></td>
+<td>
+<p>a logical argument indicating whether to sort only within each partition</p>
+</td></tr>
+<tr valign="top"><td><code>decreasing</code></td>
+<td>
+<p>a logical argument indicating sorting order for columns when
+a character vector is specified for col</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame where all elements are sorted.
+</p>
+
+
+<h3>Note</h3>
+
+<p>arrange(SparkDataFrame, Column) since 1.4.0
+</p>
+<p>arrange(SparkDataFrame, character) since 1.4.0
+</p>
+<p>orderBy(SparkDataFrame, characterOrColumn) since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D arrange(df, df$col1)
+##D arrange(df, asc(df$col1), desc(abs(df$col2)))
+##D arrange(df, &quot;col1&quot;, decreasing = TRUE)
+##D arrange(df, &quot;col1&quot;, &quot;col2&quot;, decreasing = c(TRUE, FALSE))
+##D arrange(df, &quot;col1&quot;, &quot;col2&quot;, withinPartitions = TRUE)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/as.data.frame.html b/site/docs/2.4.3/api/R/as.data.frame.html
new file mode 100644
index 0000000..cec275e
--- /dev/null
+++ b/site/docs/2.4.3/api/R/as.data.frame.html
@@ -0,0 +1,127 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Download data from a SparkDataFrame into a R data.frame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for as.data.frame {SparkR}"><tr><td>as.data.frame {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Download data from a SparkDataFrame into a R data.frame</h2>
+
+<h3>Description</h3>
+
+<p>This function downloads the contents of a SparkDataFrame into an R's data.frame.
+Since data.frames are held in memory, ensure that you have enough memory
+in your system to accommodate the contents.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+as.data.frame(x, row.names = NULL, optional = FALSE, ...)
+
+## S4 method for signature 'SparkDataFrame'
+as.data.frame(x, row.names = NULL,
+  optional = FALSE, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>row.names</code></td>
+<td>
+<p><code>NULL</code> or a character vector giving the row names for the data frame.</p>
+</td></tr>
+<tr valign="top"><td><code>optional</code></td>
+<td>
+<p>If <code>TRUE</code>, converting column names is optional.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional arguments to pass to base::as.data.frame.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A data.frame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>as.data.frame since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D irisDF &lt;- createDataFrame(iris)
+##D df &lt;- as.data.frame(irisDF[irisDF$Species == &quot;setosa&quot;, ])
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/attach.html b/site/docs/2.4.3/api/R/attach.html
new file mode 100644
index 0000000..24e8f15
--- /dev/null
+++ b/site/docs/2.4.3/api/R/attach.html
@@ -0,0 +1,123 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Attach SparkDataFrame to R search path</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for attach,SparkDataFrame-method {SparkR}"><tr><td>attach,SparkDataFrame-method {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Attach SparkDataFrame to R search path</h2>
+
+<h3>Description</h3>
+
+<p>The specified SparkDataFrame is attached to the R search path. This means that
+the SparkDataFrame is searched by R when evaluating a variable, so columns in
+the SparkDataFrame can be accessed by simply giving their names.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'SparkDataFrame'
+attach(what, pos = 2L,
+  name = deparse(substitute(what), backtick = FALSE),
+  warn.conflicts = TRUE)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>what</code></td>
+<td>
+<p>(SparkDataFrame) The SparkDataFrame to attach</p>
+</td></tr>
+<tr valign="top"><td><code>pos</code></td>
+<td>
+<p>(integer) Specify position in search() where to attach.</p>
+</td></tr>
+<tr valign="top"><td><code>name</code></td>
+<td>
+<p>(character) Name to use for the attached SparkDataFrame. Names
+starting with package: are reserved for library.</p>
+</td></tr>
+<tr valign="top"><td><code>warn.conflicts</code></td>
+<td>
+<p>(logical) If TRUE, warnings are printed about conflicts
+from attaching the database, unless that SparkDataFrame contains an object</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>attach since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="../../base/html/detach.html">detach</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D attach(irisDf)
+##D summary(Sepal_Width)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/avg.html b/site/docs/2.4.3/api/R/avg.html
new file mode 100644
index 0000000..cb021aa
--- /dev/null
+++ b/site/docs/2.4.3/api/R/avg.html
@@ -0,0 +1,67 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: avg</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for avg {SparkR}"><tr><td>avg {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>avg</h2>
+
+<h3>Description</h3>
+
+<p>Aggregate function: returns the average of the values in a group.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+avg(x, ...)
+
+## S4 method for signature 'Column'
+avg(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on or a GroupedData object.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s) when <code>x</code> is a GroupedData object.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>avg since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other aggregate functions: <code><a href="column_aggregate_functions.html">column_aggregate_functions</a></code>,
+<code><a href="corr.html">corr</a></code>, <code><a href="count.html">count</a></code>,
+<code><a href="cov.html">cov</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="last.html">last</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: avg(df$c)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/awaitTermination.html b/site/docs/2.4.3/api/R/awaitTermination.html
new file mode 100644
index 0000000..38add9b
--- /dev/null
+++ b/site/docs/2.4.3/api/R/awaitTermination.html
@@ -0,0 +1,84 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: awaitTermination</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for awaitTermination {SparkR}"><tr><td>awaitTermination {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>awaitTermination</h2>
+
+<h3>Description</h3>
+
+<p>Waits for the termination of the query, either by <code>stopQuery</code> or by an error.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+awaitTermination(x, timeout = NULL)
+
+## S4 method for signature 'StreamingQuery'
+awaitTermination(x, timeout = NULL)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a StreamingQuery.</p>
+</td></tr>
+<tr valign="top"><td><code>timeout</code></td>
+<td>
+<p>time to wait in milliseconds, if omitted, wait indefinitely until <code>stopQuery</code>
+is called or an error has occurred.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>If the query has terminated, then all subsequent calls to this method will return TRUE
+immediately.
+</p>
+
+
+<h3>Value</h3>
+
+<p>TRUE if query has terminated within the timeout period; nothing if timeout is not
+specified.
+</p>
+
+
+<h3>Note</h3>
+
+<p>awaitTermination(StreamingQuery) since 2.2.0
+</p>
+<p>experimental
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other StreamingQuery methods: <code><a href="explain.html">explain</a></code>,
+<code><a href="isActive.html">isActive</a></code>, <code><a href="lastProgress.html">lastProgress</a></code>,
+<code><a href="queryName.html">queryName</a></code>, <code><a href="status.html">status</a></code>,
+<code><a href="stopQuery.html">stopQuery</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run:  awaitTermination(sq, 10000) 
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/between.html b/site/docs/2.4.3/api/R/between.html
new file mode 100644
index 0000000..77c6838
--- /dev/null
+++ b/site/docs/2.4.3/api/R/between.html
@@ -0,0 +1,55 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: between</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for between {SparkR}"><tr><td>between {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>between</h2>
+
+<h3>Description</h3>
+
+<p>Test if the column is between the lower bound and upper bound, inclusive.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+between(x, bounds)
+
+## S4 method for signature 'Column'
+between(x, bounds)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column</p>
+</td></tr>
+<tr valign="top"><td><code>bounds</code></td>
+<td>
+<p>lower and upper bounds</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>between since 1.5.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="alias.html">alias</a></code>, <code><a href="cast.html">cast</a></code>,
+<code><a href="endsWith.html">endsWith</a></code>, <code><a href="otherwise.html">otherwise</a></code>,
+<code><a href="over.html">over</a></code>, <code><a href="startsWith.html">startsWith</a></code>,
+<code><a href="substr.html">substr</a></code>
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/broadcast.html b/site/docs/2.4.3/api/R/broadcast.html
new file mode 100644
index 0000000..9238a64
--- /dev/null
+++ b/site/docs/2.4.3/api/R/broadcast.html
@@ -0,0 +1,119 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: broadcast</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for broadcast {SparkR}"><tr><td>broadcast {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>broadcast</h2>
+
+<h3>Description</h3>
+
+<p>Return a new SparkDataFrame marked as small enough for use in broadcast joins.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+broadcast(x)
+
+## S4 method for signature 'SparkDataFrame'
+broadcast(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Equivalent to <code>hint(x, "broadcast")</code>.
+</p>
+
+
+<h3>Value</h3>
+
+<p>a SparkDataFrame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>broadcast since 2.3.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="cache.html">cache</a></code>, <code><a href="checkpoint.html">checkpoint</a></code>,
+<code><a href="coalesce.html">coalesce</a></code>, <code><a href="collect.html">collect</a></code>,
+<code><a href="columns.html">colnames</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(mtcars)
+##D avg_mpg &lt;- mean(groupBy(createDataFrame(mtcars), &quot;cyl&quot;), &quot;mpg&quot;)
+##D 
+##D head(join(df, broadcast(avg_mpg), df$cyl == avg_mpg$cyl))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/cache.html b/site/docs/2.4.3/api/R/cache.html
new file mode 100644
index 0000000..6fc4627
--- /dev/null
+++ b/site/docs/2.4.3/api/R/cache.html
@@ -0,0 +1,107 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Cache</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for cache {SparkR}"><tr><td>cache {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Cache</h2>
+
+<h3>Description</h3>
+
+<p>Persist with the default storage level (MEMORY_ONLY).
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+cache(x)
+
+## S4 method for signature 'SparkDataFrame'
+cache(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>cache since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="checkpoint.html">checkpoint</a></code>,
+<code><a href="coalesce.html">coalesce</a></code>, <code><a href="collect.html">collect</a></code>,
+<code><a href="columns.html">colnames</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D cache(df)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/cacheTable.html b/site/docs/2.4.3/api/R/cacheTable.html
new file mode 100644
index 0000000..9abfb60
--- /dev/null
+++ b/site/docs/2.4.3/api/R/cacheTable.html
@@ -0,0 +1,65 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Cache Table</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for cacheTable {SparkR}"><tr><td>cacheTable {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Cache Table</h2>
+
+<h3>Description</h3>
+
+<p>Caches the specified table in-memory.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+cacheTable(tableName)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>the qualified or unqualified name that designates a table. If no database
+identifier is provided, it refers to a table in the current database.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>cacheTable since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D createOrReplaceTempView(df, &quot;table&quot;)
+##D cacheTable(&quot;table&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/cancelJobGroup.html b/site/docs/2.4.3/api/R/cancelJobGroup.html
new file mode 100644
index 0000000..8084f19
--- /dev/null
+++ b/site/docs/2.4.3/api/R/cancelJobGroup.html
@@ -0,0 +1,55 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Cancel active jobs for the specified group</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for cancelJobGroup {SparkR}"><tr><td>cancelJobGroup {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Cancel active jobs for the specified group</h2>
+
+<h3>Description</h3>
+
+<p>Cancel active jobs for the specified group
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+cancelJobGroup(groupId)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>groupId</code></td>
+<td>
+<p>the ID of job group to be cancelled</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>cancelJobGroup since 1.5.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D cancelJobGroup(&quot;myJobGroup&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/cast.html b/site/docs/2.4.3/api/R/cast.html
new file mode 100644
index 0000000..061d0b2
--- /dev/null
+++ b/site/docs/2.4.3/api/R/cast.html
@@ -0,0 +1,72 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Casts the column to a different data type.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for cast {SparkR}"><tr><td>cast {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Casts the column to a different data type.</h2>
+
+<h3>Description</h3>
+
+<p>Casts the column to a different data type.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+cast(x, dataType)
+
+## S4 method for signature 'Column'
+cast(x, dataType)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column.</p>
+</td></tr>
+<tr valign="top"><td><code>dataType</code></td>
+<td>
+<p>a character object describing the target data type.
+See
+<a href="https://spark.apache.org/docs/latest/sparkr.html#data-type-mapping-between-r-and-spark">
+Spark Data Types</a> for available data types.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>cast since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="alias.html">alias</a></code>,
+<code><a href="between.html">between</a></code>, <code><a href="endsWith.html">endsWith</a></code>,
+<code><a href="otherwise.html">otherwise</a></code>, <code><a href="over.html">over</a></code>,
+<code><a href="startsWith.html">startsWith</a></code>, <code><a href="substr.html">substr</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D   cast(df$age, &quot;string&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/checkpoint.html b/site/docs/2.4.3/api/R/checkpoint.html
new file mode 100644
index 0000000..b3b9ef9
--- /dev/null
+++ b/site/docs/2.4.3/api/R/checkpoint.html
@@ -0,0 +1,120 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: checkpoint</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for checkpoint {SparkR}"><tr><td>checkpoint {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>checkpoint</h2>
+
+<h3>Description</h3>
+
+<p>Returns a checkpointed version of this SparkDataFrame. Checkpointing can be used to truncate the
+logical plan, which is especially useful in iterative algorithms where the plan may grow
+exponentially. It will be saved to files inside the checkpoint directory set with
+<code>setCheckpointDir</code>
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+checkpoint(x, eager = TRUE)
+
+## S4 method for signature 'SparkDataFrame'
+checkpoint(x, eager = TRUE)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+<tr valign="top"><td><code>eager</code></td>
+<td>
+<p>whether to checkpoint this SparkDataFrame immediately</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>a new checkpointed SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>checkpoint since 2.2.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="setCheckpointDir.html">setCheckpointDir</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="coalesce.html">coalesce</a></code>, <code><a href="collect.html">collect</a></code>,
+<code><a href="columns.html">colnames</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D setCheckpointDir(&quot;/checkpoint&quot;)
+##D df &lt;- checkpoint(df)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/clearCache.html b/site/docs/2.4.3/api/R/clearCache.html
new file mode 100644
index 0000000..dac3b33
--- /dev/null
+++ b/site/docs/2.4.3/api/R/clearCache.html
@@ -0,0 +1,44 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Clear Cache</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for clearCache {SparkR}"><tr><td>clearCache {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Clear Cache</h2>
+
+<h3>Description</h3>
+
+<p>Removes all cached tables from the in-memory cache.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+clearCache()
+</pre>
+
+
+<h3>Note</h3>
+
+<p>clearCache since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D clearCache()
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/clearJobGroup.html b/site/docs/2.4.3/api/R/clearJobGroup.html
new file mode 100644
index 0000000..cdd0fd5
--- /dev/null
+++ b/site/docs/2.4.3/api/R/clearJobGroup.html
@@ -0,0 +1,45 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Clear current job group ID and its description</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for clearJobGroup {SparkR}"><tr><td>clearJobGroup {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Clear current job group ID and its description</h2>
+
+<h3>Description</h3>
+
+<p>Clear current job group ID and its description
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+clearJobGroup()
+</pre>
+
+
+<h3>Note</h3>
+
+<p>clearJobGroup since 1.5.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D clearJobGroup()
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/coalesce.html b/site/docs/2.4.3/api/R/coalesce.html
new file mode 100644
index 0000000..b499ab1
--- /dev/null
+++ b/site/docs/2.4.3/api/R/coalesce.html
@@ -0,0 +1,132 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Coalesce</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for coalesce {SparkR}"><tr><td>coalesce {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Coalesce</h2>
+
+<h3>Description</h3>
+
+<p>Returns a new SparkDataFrame that has exactly <code>numPartitions</code> partitions.
+This operation results in a narrow dependency, e.g. if you go from 1000 partitions to 100
+partitions, there will not be a shuffle, instead each of the 100 new partitions will claim 10 of
+the current partitions. If a larger number of partitions is requested, it will stay at the
+current number of partitions.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+coalesce(x, ...)
+
+## S4 method for signature 'SparkDataFrame'
+coalesce(x, numPartitions)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s).</p>
+</td></tr>
+<tr valign="top"><td><code>numPartitions</code></td>
+<td>
+<p>the number of partitions to use.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>However, if you're doing a drastic coalesce on a SparkDataFrame, e.g. to numPartitions = 1,
+this may result in your computation taking place on fewer nodes than
+you like (e.g. one node in the case of numPartitions = 1). To avoid this,
+call <code>repartition</code>. This will add a shuffle step, but means the
+current upstream partitions will be executed in parallel (per whatever
+the current partitioning is).
+</p>
+
+
+<h3>Note</h3>
+
+<p>coalesce(SparkDataFrame) since 2.1.1
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="repartition.html">repartition</a>, <a href="repartitionByRange.html">repartitionByRange</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="collect.html">collect</a></code>,
+<code><a href="columns.html">colnames</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D newDF &lt;- coalesce(df, 1L)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/collect.html b/site/docs/2.4.3/api/R/collect.html
new file mode 100644
index 0000000..6bfbeef
--- /dev/null
+++ b/site/docs/2.4.3/api/R/collect.html
@@ -0,0 +1,118 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Collects all the elements of a SparkDataFrame and coerces...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for collect {SparkR}"><tr><td>collect {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Collects all the elements of a SparkDataFrame and coerces them into an R data.frame.</h2>
+
+<h3>Description</h3>
+
+<p>Collects all the elements of a SparkDataFrame and coerces them into an R data.frame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+collect(x, ...)
+
+## S4 method for signature 'SparkDataFrame'
+collect(x, stringsAsFactors = FALSE)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>further arguments to be passed to or from other methods.</p>
+</td></tr>
+<tr valign="top"><td><code>stringsAsFactors</code></td>
+<td>
+<p>(Optional) a logical indicating whether or not string columns
+should be converted to factors. FALSE by default.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>collect since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="columns.html">colnames</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D collected &lt;- collect(df)
+##D class(collected)
+##D firstName &lt;- names(collected)[1]
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/coltypes.html b/site/docs/2.4.3/api/R/coltypes.html
new file mode 100644
index 0000000..74159e9
--- /dev/null
+++ b/site/docs/2.4.3/api/R/coltypes.html
@@ -0,0 +1,133 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: coltypes</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for coltypes {SparkR}"><tr><td>coltypes {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>coltypes</h2>
+
+<h3>Description</h3>
+
+<p>Get column types of a SparkDataFrame
+</p>
+<p>Set the column types of a SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+coltypes(x)
+
+coltypes(x) &lt;- value
+
+## S4 method for signature 'SparkDataFrame'
+coltypes(x)
+
+## S4 replacement method for signature 'SparkDataFrame,character'
+coltypes(x) &lt;- value
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+<tr valign="top"><td><code>value</code></td>
+<td>
+<p>A character vector with the target column types for the given
+SparkDataFrame. Column types can be one of integer, numeric/double, character, logical, or NA
+to keep that column as-is.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>value A character vector with the column types of the given SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>coltypes since 1.6.0
+</p>
+<p>coltypes&lt;- since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D irisDF &lt;- createDataFrame(iris)
+##D coltypes(irisDF) # get column types
+## End(Not run)
+## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D coltypes(df) &lt;- c(&quot;character&quot;, &quot;integer&quot;) # set column types
+##D coltypes(df) &lt;- c(NA, &quot;numeric&quot;) # set column types
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column.html b/site/docs/2.4.3/api/R/column.html
new file mode 100644
index 0000000..2b29edb
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column.html
@@ -0,0 +1,77 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: S4 class that represents a SparkDataFrame column</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column {SparkR}"><tr><td>column {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>S4 class that represents a SparkDataFrame column</h2>
+
+<h3>Description</h3>
+
+<p>The column class supports unary, binary operations on SparkDataFrame columns
+</p>
+<p>Returns a Column based on the given column name.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+column(x)
+
+## S4 method for signature 'jobj'
+column(x)
+
+## S4 method for signature 'character'
+column(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Character column name.</p>
+</td></tr>
+</table>
+
+
+<h3>Slots</h3>
+
+
+<dl>
+<dt><code>jc</code></dt><dd><p>reference to JVM SparkDataFrame column</p>
+</dd>
+</dl>
+
+
+<h3>Note</h3>
+
+<p>Column since 1.4.0
+</p>
+<p>column since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other non-aggregate functions: <code><a href="column_nonaggregate_functions.html">column_nonaggregate_functions</a></code>,
+<code><a href="not.html">not</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: column(&quot;name&quot;)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_aggregate_functions.html b/site/docs/2.4.3/api/R/column_aggregate_functions.html
new file mode 100644
index 0000000..a2f7e52
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_aggregate_functions.html
@@ -0,0 +1,352 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Aggregate functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_aggregate_functions {SparkR}"><tr><td>column_aggregate_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Aggregate functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Aggregate functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+approxCountDistinct(x, ...)
+
+collect_list(x)
+
+collect_set(x)
+
+countDistinct(x, ...)
+
+grouping_bit(x)
+
+grouping_id(x, ...)
+
+kurtosis(x)
+
+n_distinct(x, ...)
+
+sd(x, na.rm = FALSE)
+
+skewness(x)
+
+stddev(x)
+
+stddev_pop(x)
+
+stddev_samp(x)
+
+sumDistinct(x)
+
+var(x, y = NULL, na.rm = FALSE, use)
+
+variance(x)
+
+var_pop(x)
+
+var_samp(x)
+
+## S4 method for signature 'Column'
+approxCountDistinct(x, rsd = 0.05)
+
+## S4 method for signature 'Column'
+kurtosis(x)
+
+## S4 method for signature 'Column'
+max(x)
+
+## S4 method for signature 'Column'
+mean(x)
+
+## S4 method for signature 'Column'
+min(x)
+
+## S4 method for signature 'Column'
+sd(x)
+
+## S4 method for signature 'Column'
+skewness(x)
+
+## S4 method for signature 'Column'
+stddev(x)
+
+## S4 method for signature 'Column'
+stddev_pop(x)
+
+## S4 method for signature 'Column'
+stddev_samp(x)
+
+## S4 method for signature 'Column'
+sum(x)
+
+## S4 method for signature 'Column'
+sumDistinct(x)
+
+## S4 method for signature 'Column'
+var(x)
+
+## S4 method for signature 'Column'
+variance(x)
+
+## S4 method for signature 'Column'
+var_pop(x)
+
+## S4 method for signature 'Column'
+var_samp(x)
+
+## S4 method for signature 'Column'
+approxCountDistinct(x, rsd = 0.05)
+
+## S4 method for signature 'Column'
+countDistinct(x, ...)
+
+## S4 method for signature 'Column'
+n_distinct(x, ...)
+
+## S4 method for signature 'Column'
+collect_list(x)
+
+## S4 method for signature 'Column'
+collect_set(x)
+
+## S4 method for signature 'Column'
+grouping_bit(x)
+
+## S4 method for signature 'Column'
+grouping_id(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s). For example, it could be used to pass additional Columns.</p>
+</td></tr>
+<tr valign="top"><td><code>y, na.rm, use</code></td>
+<td>
+<p>currently not used.</p>
+</td></tr>
+<tr valign="top"><td><code>rsd</code></td>
+<td>
+<p>maximum estimation error allowed (default = 0.05).</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>approxCountDistinct</code>: Returns the approximate number of distinct items in a group.
+</p>
+<p><code>kurtosis</code>: Returns the kurtosis of the values in a group.
+</p>
+<p><code>max</code>: Returns the maximum value of the expression in a group.
+</p>
+<p><code>mean</code>: Returns the average of the values in a group. Alias for <code>avg</code>.
+</p>
+<p><code>min</code>: Returns the minimum value of the expression in a group.
+</p>
+<p><code>sd</code>: Alias for <code>stddev_samp</code>.
+</p>
+<p><code>skewness</code>: Returns the skewness of the values in a group.
+</p>
+<p><code>stddev</code>: Alias for <code>std_dev</code>.
+</p>
+<p><code>stddev_pop</code>: Returns the population standard deviation of the expression in a group.
+</p>
+<p><code>stddev_samp</code>: Returns the unbiased sample standard deviation of the expression in a group.
+</p>
+<p><code>sum</code>: Returns the sum of all values in the expression.
+</p>
+<p><code>sumDistinct</code>: Returns the sum of distinct values in the expression.
+</p>
+<p><code>var</code>: Alias for <code>var_samp</code>.
+</p>
+<p><code>var_pop</code>: Returns the population variance of the values in a group.
+</p>
+<p><code>var_samp</code>: Returns the unbiased variance of the values in a group.
+</p>
+<p><code>countDistinct</code>: Returns the number of distinct items in a group.
+</p>
+<p><code>n_distinct</code>: Returns the number of distinct items in a group.
+</p>
+<p><code>collect_list</code>: Creates a list of objects with duplicates.
+Note: the function is non-deterministic because the order of collected results depends
+on order of rows which may be non-deterministic after a shuffle.
+</p>
+<p><code>collect_set</code>: Creates a list of objects with duplicate elements eliminated.
+Note: the function is non-deterministic because the order of collected results depends
+on order of rows which may be non-deterministic after a shuffle.
+</p>
+<p><code>grouping_bit</code>: Indicates whether a specified column in a GROUP BY list is aggregated or
+not, returns 1 for aggregated or 0 for not aggregated in the result set. Same as <code>GROUPING</code>
+in SQL and <code>grouping</code> function in Scala.
+</p>
+<p><code>grouping_id</code>: Returns the level of grouping.
+Equals to <code>
+grouping_bit(c1) * 2^(n - 1) + grouping_bit(c2) * 2^(n - 2)  + ... + grouping_bit(cn)
+</code>.
+</p>
+
+
+<h3>Note</h3>
+
+<p>approxCountDistinct(Column) since 1.4.0
+</p>
+<p>kurtosis since 1.6.0
+</p>
+<p>max since 1.5.0
+</p>
+<p>mean since 1.5.0
+</p>
+<p>min since 1.5.0
+</p>
+<p>sd since 1.6.0
+</p>
+<p>skewness since 1.6.0
+</p>
+<p>stddev since 1.6.0
+</p>
+<p>stddev_pop since 1.6.0
+</p>
+<p>stddev_samp since 1.6.0
+</p>
+<p>sum since 1.5.0
+</p>
+<p>sumDistinct since 1.4.0
+</p>
+<p>var since 1.6.0
+</p>
+<p>variance since 1.6.0
+</p>
+<p>var_pop since 1.5.0
+</p>
+<p>var_samp since 1.6.0
+</p>
+<p>approxCountDistinct(Column, numeric) since 1.4.0
+</p>
+<p>countDistinct since 1.4.0
+</p>
+<p>n_distinct since 1.4.0
+</p>
+<p>collect_list since 2.3.0
+</p>
+<p>collect_set since 2.3.0
+</p>
+<p>grouping_bit since 2.3.0
+</p>
+<p>grouping_id since 2.3.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other aggregate functions: <code><a href="avg.html">avg</a></code>,
+<code><a href="corr.html">corr</a></code>, <code><a href="count.html">count</a></code>,
+<code><a href="cov.html">cov</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="last.html">last</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, approxCountDistinct(df$gear)))
+##D head(select(df, approxCountDistinct(df$gear, 0.02)))
+##D head(select(df, countDistinct(df$gear, df$cyl)))
+##D head(select(df, n_distinct(df$gear)))
+##D head(distinct(select(df, &quot;gear&quot;)))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, mean(df$mpg), sd(df$mpg), skewness(df$mpg), kurtosis(df$mpg)))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, avg(df$mpg), mean(df$mpg), sum(df$mpg), min(df$wt), max(df$qsec)))
+##D 
+##D # metrics by num of cylinders
+##D tmp &lt;- agg(groupBy(df, &quot;cyl&quot;), avg(df$mpg), avg(df$hp), avg(df$wt), avg(df$qsec))
+##D head(orderBy(tmp, &quot;cyl&quot;))
+##D 
+##D # car with the max mpg
+##D mpg_max &lt;- as.numeric(collect(agg(df, max(df$mpg))))
+##D head(where(df, df$mpg == mpg_max))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, sd(df$mpg), stddev(df$mpg), stddev_pop(df$wt), stddev_samp(df$qsec)))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, sumDistinct(df$gear)))
+##D head(distinct(select(df, &quot;gear&quot;)))
+## End(Not run)
+
+## Not run: 
+##D head(agg(df, var(df$mpg), variance(df$mpg), var_pop(df$mpg), var_samp(df$mpg)))
+## End(Not run)
+
+## Not run: 
+##D df2 = df[df$mpg &gt; 20, ]
+##D collect(select(df2, collect_list(df2$gear)))
+##D collect(select(df2, collect_set(df2$gear)))
+## End(Not run)
+
+## Not run: 
+##D # With cube
+##D agg(
+##D   cube(df, &quot;cyl&quot;, &quot;gear&quot;, &quot;am&quot;),
+##D   mean(df$mpg),
+##D   grouping_bit(df$cyl), grouping_bit(df$gear), grouping_bit(df$am)
+##D )
+##D 
+##D # With rollup
+##D agg(
+##D   rollup(df, &quot;cyl&quot;, &quot;gear&quot;, &quot;am&quot;),
+##D   mean(df$mpg),
+##D   grouping_bit(df$cyl), grouping_bit(df$gear), grouping_bit(df$am)
+##D )
+## End(Not run)
+
+## Not run: 
+##D # With cube
+##D agg(
+##D   cube(df, &quot;cyl&quot;, &quot;gear&quot;, &quot;am&quot;),
+##D   mean(df$mpg),
+##D   grouping_id(df$cyl, df$gear, df$am)
+##D )
+##D 
+##D # With rollup
+##D agg(
+##D   rollup(df, &quot;cyl&quot;, &quot;gear&quot;, &quot;am&quot;),
+##D   mean(df$mpg),
+##D   grouping_id(df$cyl, df$gear, df$am)
+##D )
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_collection_functions.html b/site/docs/2.4.3/api/R/column_collection_functions.html
new file mode 100644
index 0000000..4915d72
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_collection_functions.html
@@ -0,0 +1,503 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Collection functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_collection_functions {SparkR}"><tr><td>column_collection_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Collection functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Collection functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+array_contains(x, value)
+
+array_distinct(x)
+
+array_except(x, y)
+
+array_intersect(x, y)
+
+array_join(x, delimiter, ...)
+
+array_max(x)
+
+array_min(x)
+
+array_position(x, value)
+
+array_remove(x, value)
+
+array_repeat(x, count)
+
+array_sort(x)
+
+arrays_overlap(x, y)
+
+array_union(x, y)
+
+arrays_zip(x, ...)
+
+concat(x, ...)
+
+element_at(x, extraction)
+
+explode(x)
+
+explode_outer(x)
+
+flatten(x)
+
+from_json(x, schema, ...)
+
+map_from_arrays(x, y)
+
+map_keys(x)
+
+map_values(x)
+
+posexplode(x)
+
+posexplode_outer(x)
+
+reverse(x)
+
+shuffle(x)
+
+size(x)
+
+slice(x, start, length)
+
+sort_array(x, asc = TRUE)
+
+to_json(x, ...)
+
+## S4 method for signature 'Column'
+reverse(x)
+
+## S4 method for signature 'Column'
+to_json(x, ...)
+
+## S4 method for signature 'Column'
+concat(x, ...)
+
+## S4 method for signature 'Column,characterOrstructType'
+from_json(x, schema,
+  as.json.array = FALSE, ...)
+
+## S4 method for signature 'Column'
+array_contains(x, value)
+
+## S4 method for signature 'Column'
+array_distinct(x)
+
+## S4 method for signature 'Column,Column'
+array_except(x, y)
+
+## S4 method for signature 'Column,Column'
+array_intersect(x, y)
+
+## S4 method for signature 'Column,character'
+array_join(x, delimiter,
+  nullReplacement = NULL)
+
+## S4 method for signature 'Column'
+array_max(x)
+
+## S4 method for signature 'Column'
+array_min(x)
+
+## S4 method for signature 'Column'
+array_position(x, value)
+
+## S4 method for signature 'Column'
+array_remove(x, value)
+
+## S4 method for signature 'Column,numericOrColumn'
+array_repeat(x, count)
+
+## S4 method for signature 'Column'
+array_sort(x)
+
+## S4 method for signature 'Column,Column'
+arrays_overlap(x, y)
+
+## S4 method for signature 'Column,Column'
+array_union(x, y)
+
+## S4 method for signature 'Column'
+arrays_zip(x, ...)
+
+## S4 method for signature 'Column'
+shuffle(x)
+
+## S4 method for signature 'Column'
+flatten(x)
+
+## S4 method for signature 'Column,Column'
+map_from_arrays(x, y)
+
+## S4 method for signature 'Column'
+map_keys(x)
+
+## S4 method for signature 'Column'
+map_values(x)
+
+## S4 method for signature 'Column'
+element_at(x, extraction)
+
+## S4 method for signature 'Column'
+explode(x)
+
+## S4 method for signature 'Column'
+size(x)
+
+## S4 method for signature 'Column'
+slice(x, start, length)
+
+## S4 method for signature 'Column'
+sort_array(x, asc = TRUE)
+
+## S4 method for signature 'Column'
+posexplode(x)
+
+## S4 method for signature 'Column'
+explode_outer(x)
+
+## S4 method for signature 'Column'
+posexplode_outer(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on. Note the difference in the following methods:
+</p>
+
+<ul>
+<li> <p><code>to_json</code>: it is the column containing the struct, array of the structs,
+the map or array of maps.
+</p>
+</li>
+<li> <p><code>from_json</code>: it is the column containing the JSON string.
+</p>
+</li></ul>
+</td></tr>
+<tr valign="top"><td><code>value</code></td>
+<td>
+<p>A value to compute on.
+</p>
+
+<ul>
+<li> <p><code>array_contains</code>: a value to be checked if contained in the column.
+</p>
+</li>
+<li> <p><code>array_position</code>: a value to locate in the given array.
+</p>
+</li>
+<li> <p><code>array_remove</code>: a value to remove in the given array.
+</p>
+</li></ul>
+</td></tr>
+<tr valign="top"><td><code>y</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>delimiter</code></td>
+<td>
+<p>a character string that is used to concatenate the elements of column.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s). In <code>to_json</code> and <code>from_json</code>, this contains
+additional named properties to control how it is converted, accepts the same
+options as the JSON data source.  In <code>arrays_zip</code>, this contains additional
+Columns of arrays to be merged.</p>
+</td></tr>
+<tr valign="top"><td><code>count</code></td>
+<td>
+<p>a Column or constant determining the number of repetitions.</p>
+</td></tr>
+<tr valign="top"><td><code>extraction</code></td>
+<td>
+<p>index to check for in array or key to check for in map</p>
+</td></tr>
+<tr valign="top"><td><code>schema</code></td>
+<td>
+<p>a structType object to use as the schema to use when parsing the JSON string.
+Since Spark 2.3, the DDL-formatted string is also supported for the schema.</p>
+</td></tr>
+<tr valign="top"><td><code>start</code></td>
+<td>
+<p>an index indicating the first element occurring in the result.</p>
+</td></tr>
+<tr valign="top"><td><code>length</code></td>
+<td>
+<p>a number of consecutive elements chosen to the result.</p>
+</td></tr>
+<tr valign="top"><td><code>asc</code></td>
+<td>
+<p>a logical flag indicating the sorting order.
+TRUE, sorting is in ascending order.
+FALSE, sorting is in descending order.</p>
+</td></tr>
+<tr valign="top"><td><code>as.json.array</code></td>
+<td>
+<p>indicating if input string is JSON array of objects or a single object.</p>
+</td></tr>
+<tr valign="top"><td><code>nullReplacement</code></td>
+<td>
+<p>an optional character string that is used to replace the Null values.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>reverse</code>: Returns a reversed string or an array with reverse order of elements.
+</p>
+<p><code>to_json</code>: Converts a column containing a <code>structType</code>, a <code>mapType</code>
+or an <code>arrayType</code> into a Column of JSON string.
+Resolving the Column can fail if an unsupported type is encountered.
+</p>
+<p><code>concat</code>: Concatenates multiple input columns together into a single column.
+The function works with strings, binary and compatible array columns.
+</p>
+<p><code>from_json</code>: Parses a column containing a JSON string into a Column of <code>structType</code>
+with the specified <code>schema</code> or array of <code>structType</code> if <code>as.json.array</code> is set
+to <code>TRUE</code>. If the string is unparseable, the Column will contain the value NA.
+</p>
+<p><code>array_contains</code>: Returns null if the array is null, true if the array contains
+the value, and false otherwise.
+</p>
+<p><code>array_distinct</code>: Removes duplicate values from the array.
+</p>
+<p><code>array_except</code>: Returns an array of the elements in the first array but not in the second
+array, without duplicates. The order of elements in the result is not determined.
+</p>
+<p><code>array_intersect</code>: Returns an array of the elements in the intersection of the given two
+arrays, without duplicates.
+</p>
+<p><code>array_join</code>: Concatenates the elements of column using the delimiter.
+Null values are replaced with nullReplacement if set, otherwise they are ignored.
+</p>
+<p><code>array_max</code>: Returns the maximum value of the array.
+</p>
+<p><code>array_min</code>: Returns the minimum value of the array.
+</p>
+<p><code>array_position</code>: Locates the position of the first occurrence of the given value
+in the given array. Returns NA if either of the arguments are NA.
+Note: The position is not zero based, but 1 based index. Returns 0 if the given
+value could not be found in the array.
+</p>
+<p><code>array_remove</code>: Removes all elements that equal to element from the given array.
+</p>
+<p><code>array_repeat</code>: Creates an array containing <code>x</code> repeated the number of times
+given by <code>count</code>.
+</p>
+<p><code>array_sort</code>: Sorts the input array in ascending order. The elements of the input array
+must be orderable. NA elements will be placed at the end of the returned array.
+</p>
+<p><code>arrays_overlap</code>: Returns true if the input arrays have at least one non-null element in
+common. If not and both arrays are non-empty and any of them contains a null, it returns null.
+It returns false otherwise.
+</p>
+<p><code>array_union</code>: Returns an array of the elements in the union of the given two arrays,
+without duplicates.
+</p>
+<p><code>arrays_zip</code>: Returns a merged array of structs in which the N-th struct contains all N-th
+values of input arrays.
+</p>
+<p><code>shuffle</code>: Returns a random permutation of the given array.
+</p>
+<p><code>flatten</code>: Creates a single array from an array of arrays.
+If a structure of nested arrays is deeper than two levels, only one level of nesting is removed.
+</p>
+<p><code>map_from_arrays</code>: Creates a new map column. The array in the first column is used for
+keys. The array in the second column is used for values. All elements in the array for key
+should not be null.
+</p>
+<p><code>map_keys</code>: Returns an unordered array containing the keys of the map.
+</p>
+<p><code>map_values</code>: Returns an unordered array containing the values of the map.
+</p>
+<p><code>element_at</code>: Returns element of array at given index in <code>extraction</code> if
+<code>x</code> is array. Returns value for the given key in <code>extraction</code> if <code>x</code> is map.
+Note: The position is not zero based, but 1 based index.
+</p>
+<p><code>explode</code>: Creates a new row for each element in the given array or map column.
+</p>
+<p><code>size</code>: Returns length of array or map.
+</p>
+<p><code>slice</code>: Returns an array containing all the elements in x from the index start
+(or starting from the end if start is negative) with the specified length.
+</p>
+<p><code>sort_array</code>: Sorts the input array in ascending or descending order according to
+the natural ordering of the array elements. NA elements will be placed at the beginning of
+the returned array in ascending order or at the end of the returned array in descending order.
+</p>
+<p><code>posexplode</code>: Creates a new row for each element with position in the given array
+or map column.
+</p>
+<p><code>explode</code>: Creates a new row for each element in the given array or map column.
+Unlike <code>explode</code>, if the array/map is <code>null</code> or empty
+then <code>null</code> is produced.
+</p>
+<p><code>posexplode_outer</code>: Creates a new row for each element with position in the given
+array or map column. Unlike <code>posexplode</code>, if the array/map is <code>null</code> or empty
+then the row (<code>null</code>, <code>null</code>) is produced.
+</p>
+
+
+<h3>Note</h3>
+
+<p>reverse since 1.5.0
+</p>
+<p>to_json since 2.2.0
+</p>
+<p>concat since 1.5.0
+</p>
+<p>from_json since 2.2.0
+</p>
+<p>array_contains since 1.6.0
+</p>
+<p>array_distinct since 2.4.0
+</p>
+<p>array_except since 2.4.0
+</p>
+<p>array_intersect since 2.4.0
+</p>
+<p>array_join since 2.4.0
+</p>
+<p>array_max since 2.4.0
+</p>
+<p>array_min since 2.4.0
+</p>
+<p>array_position since 2.4.0
+</p>
+<p>array_remove since 2.4.0
+</p>
+<p>array_repeat since 2.4.0
+</p>
+<p>array_sort since 2.4.0
+</p>
+<p>arrays_overlap since 2.4.0
+</p>
+<p>array_union since 2.4.0
+</p>
+<p>arrays_zip since 2.4.0
+</p>
+<p>shuffle since 2.4.0
+</p>
+<p>flatten since 2.4.0
+</p>
+<p>map_from_arrays since 2.4.0
+</p>
+<p>map_keys since 2.3.0
+</p>
+<p>map_values since 2.3.0
+</p>
+<p>element_at since 2.4.0
+</p>
+<p>explode since 1.5.0
+</p>
+<p>size since 1.5.0
+</p>
+<p>slice since 2.4.0
+</p>
+<p>sort_array since 1.6.0
+</p>
+<p>posexplode since 2.1.0
+</p>
+<p>explode_outer since 2.3.0
+</p>
+<p>posexplode_outer since 2.3.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+##D tmp &lt;- mutate(df, v1 = create_array(df$mpg, df$cyl, df$hp))
+##D head(select(tmp, array_contains(tmp$v1, 21), size(tmp$v1), shuffle(tmp$v1)))
+##D head(select(tmp, array_max(tmp$v1), array_min(tmp$v1), array_distinct(tmp$v1)))
+##D head(select(tmp, array_position(tmp$v1, 21), array_repeat(df$mpg, 3), array_sort(tmp$v1)))
+##D head(select(tmp, flatten(tmp$v1), reverse(tmp$v1), array_remove(tmp$v1, 21)))
+##D tmp2 &lt;- mutate(tmp, v2 = explode(tmp$v1))
+##D head(tmp2)
+##D head(select(tmp, posexplode(tmp$v1)))
+##D head(select(tmp, slice(tmp$v1, 2L, 2L)))
+##D head(select(tmp, sort_array(tmp$v1)))
+##D head(select(tmp, sort_array(tmp$v1, asc = FALSE)))
+##D tmp3 &lt;- mutate(df, v3 = create_map(df$model, df$cyl))
+##D head(select(tmp3, map_keys(tmp3$v3), map_values(tmp3$v3)))
+##D head(select(tmp3, element_at(tmp3$v3, &quot;Valiant&quot;)))
+##D tmp4 &lt;- mutate(df, v4 = create_array(df$mpg, df$cyl), v5 = create_array(df$cyl, df$hp))
+##D head(select(tmp4, concat(tmp4$v4, tmp4$v5), arrays_overlap(tmp4$v4, tmp4$v5)))
+##D head(select(tmp4, array_except(tmp4$v4, tmp4$v5), array_intersect(tmp4$v4, tmp4$v5)))
+##D head(select(tmp4, array_union(tmp4$v4, tmp4$v5)))
+##D head(select(tmp4, arrays_zip(tmp4$v4, tmp4$v5), map_from_arrays(tmp4$v4, tmp4$v5)))
+##D head(select(tmp, concat(df$mpg, df$cyl, df$hp)))
+##D tmp5 &lt;- mutate(df, v6 = create_array(df$model, df$model))
+##D head(select(tmp5, array_join(tmp5$v6, &quot;#&quot;), array_join(tmp5$v6, &quot;#&quot;, &quot;NULL&quot;)))
+## End(Not run)
+
+## Not run: 
+##D # Converts a struct into a JSON object
+##D df2 &lt;- sql(&quot;SELECT named_struct(&#39;date&#39;, cast(&#39;2000-01-01&#39; as date)) as d&quot;)
+##D select(df2, to_json(df2$d, dateFormat = &#39;dd/MM/yyyy&#39;))
+##D 
+##D # Converts an array of structs into a JSON array
+##D df2 &lt;- sql(&quot;SELECT array(named_struct(&#39;name&#39;, &#39;Bob&#39;), named_struct(&#39;name&#39;, &#39;Alice&#39;)) as people&quot;)
+##D df2 &lt;- mutate(df2, people_json = to_json(df2$people))
+##D 
+##D # Converts a map into a JSON object
+##D df2 &lt;- sql(&quot;SELECT map(&#39;name&#39;, &#39;Bob&#39;)) as people&quot;)
+##D df2 &lt;- mutate(df2, people_json = to_json(df2$people))
+##D 
+##D # Converts an array of maps into a JSON array
+##D df2 &lt;- sql(&quot;SELECT array(map(&#39;name&#39;, &#39;Bob&#39;), map(&#39;name&#39;, &#39;Alice&#39;)) as people&quot;)
+##D df2 &lt;- mutate(df2, people_json = to_json(df2$people))
+## End(Not run)
+
+## Not run: 
+##D df2 &lt;- sql(&quot;SELECT named_struct(&#39;date&#39;, cast(&#39;2000-01-01&#39; as date)) as d&quot;)
+##D df2 &lt;- mutate(df2, d2 = to_json(df2$d, dateFormat = &#39;dd/MM/yyyy&#39;))
+##D schema &lt;- structType(structField(&quot;date&quot;, &quot;string&quot;))
+##D head(select(df2, from_json(df2$d2, schema, dateFormat = &#39;dd/MM/yyyy&#39;)))
+##D df2 &lt;- sql(&quot;SELECT named_struct(&#39;name&#39;, &#39;Bob&#39;) as people&quot;)
+##D df2 &lt;- mutate(df2, people_json = to_json(df2$people))
+##D schema &lt;- structType(structField(&quot;name&quot;, &quot;string&quot;))
+##D head(select(df2, from_json(df2$people_json, schema)))
+##D head(select(df2, from_json(df2$people_json, &quot;name STRING&quot;)))
+## End(Not run)
+
+## Not run: 
+##D df2 &lt;- createDataFrame(data.frame(
+##D   id = c(1, 2, 3), text = c(&quot;a,b,c&quot;, NA, &quot;d,e&quot;)
+##D ))
+##D 
+##D head(select(df2, df2$id, explode_outer(split_string(df2$text, &quot;,&quot;))))
+##D head(select(df2, df2$id, posexplode_outer(split_string(df2$text, &quot;,&quot;))))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_datetime_diff_functions.html b/site/docs/2.4.3/api/R/column_datetime_diff_functions.html
new file mode 100644
index 0000000..f7705f2
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_datetime_diff_functions.html
@@ -0,0 +1,217 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Date time arithmetic functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_datetime_diff_functions {SparkR}"><tr><td>column_datetime_diff_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Date time arithmetic functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Date time arithmetic functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+add_months(y, x)
+
+datediff(y, x)
+
+date_add(y, x)
+
+date_format(y, x)
+
+date_sub(y, x)
+
+from_utc_timestamp(y, x)
+
+months_between(y, x)
+
+next_day(y, x)
+
+to_utc_timestamp(y, x)
+
+## S4 method for signature 'Column'
+datediff(y, x)
+
+## S4 method for signature 'Column'
+months_between(y, x)
+
+## S4 method for signature 'Column,character'
+date_format(y, x)
+
+## S4 method for signature 'Column,character'
+from_utc_timestamp(y, x)
+
+## S4 method for signature 'Column,character'
+next_day(y, x)
+
+## S4 method for signature 'Column,character'
+to_utc_timestamp(y, x)
+
+## S4 method for signature 'Column,numeric'
+add_months(y, x)
+
+## S4 method for signature 'Column,numeric'
+date_add(y, x)
+
+## S4 method for signature 'Column,numeric'
+date_sub(y, x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>y</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>For class <code>Column</code>, it is the column used to perform arithmetic operations
+with column <code>y</code>. For class <code>numeric</code>, it is the number of months or
+days to be added to or subtracted from <code>y</code>. For class <code>character</code>, it is
+</p>
+
+<ul>
+<li> <p><code>date_format</code>: date format specification.
+</p>
+</li>
+<li> <p><code>from_utc_timestamp</code>, <code>to_utc_timestamp</code>: time zone to use.
+</p>
+</li>
+<li> <p><code>next_day</code>: day of the week string.
+</p>
+</li></ul>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>datediff</code>: Returns the number of days from <code>y</code> to <code>x</code>.
+If <code>y</code> is later than <code>x</code> then the result is positive.
+</p>
+<p><code>months_between</code>: Returns number of months between dates <code>y</code> and <code>x</code>.
+If <code>y</code> is later than <code>x</code>, then the result is positive. If <code>y</code> and <code>x</code>
+are on the same day of month, or both are the last day of month, time of day will be ignored.
+Otherwise, the difference is calculated based on 31 days per month, and rounded to 8 digits.
+</p>
+<p><code>date_format</code>: Converts a date/timestamp/string to a value of string in the format
+specified by the date format given by the second argument. A pattern could be for instance
+<code>dd.MM.yyyy</code> and could return a string like '18.03.1993'. All
+pattern letters of <code>java.text.SimpleDateFormat</code> can be used.
+Note: Use when ever possible specialized functions like <code>year</code>. These benefit from a
+specialized implementation.
+</p>
+<p><code>from_utc_timestamp</code>: This is a common function for databases supporting TIMESTAMP WITHOUT
+TIMEZONE. This function takes a timestamp which is timezone-agnostic, and interprets it as a
+timestamp in UTC, and renders that timestamp as a timestamp in the given time zone.
+However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
+timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
+the given timezone.
+This function may return confusing result if the input is a string with timezone, e.g.
+(<code>2018-03-13T06:18:23+00:00</code>). The reason is that, Spark firstly cast the string to
+timestamp according to the timezone in the string, and finally display the result by converting
+the timestamp to string according to the session local timezone.
+</p>
+<p><code>next_day</code>: Given a date column, returns the first date which is later than the value of
+the date column that is on the specified day of the week. For example,
+<code>next_day("2015-07-27", "Sunday")</code> returns 2015-08-02 because that is the first Sunday
+after 2015-07-27. Day of the week parameter is case insensitive, and accepts first three or
+two characters: &quot;Mon&quot;, &quot;Tue&quot;, &quot;Wed&quot;, &quot;Thu&quot;, &quot;Fri&quot;, &quot;Sat&quot;, &quot;Sun&quot;.
+</p>
+<p><code>to_utc_timestamp</code>: This is a common function for databases supporting TIMESTAMP WITHOUT
+TIMEZONE. This function takes a timestamp which is timezone-agnostic, and interprets it as a
+timestamp in the given timezone, and renders that timestamp as a timestamp in UTC.
+However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
+timezone-agnostic. So in Spark this function just shift the timestamp value from the given
+timezone to UTC timezone.
+This function may return confusing result if the input is a string with timezone, e.g.
+(<code>2018-03-13T06:18:23+00:00</code>). The reason is that, Spark firstly cast the string to
+timestamp according to the timezone in the string, and finally display the result by converting
+the timestamp to string according to the session local timezone.
+</p>
+<p><code>add_months</code>: Returns the date that is numMonths (<code>x</code>) after startDate (<code>y</code>).
+</p>
+<p><code>date_add</code>: Returns the date that is <code>x</code> days after.
+</p>
+<p><code>date_sub</code>: Returns the date that is <code>x</code> days before.
+</p>
+
+
+<h3>Note</h3>
+
+<p>datediff since 1.5.0
+</p>
+<p>months_between since 1.5.0
+</p>
+<p>date_format since 1.5.0
+</p>
+<p>from_utc_timestamp since 1.5.0
+</p>
+<p>next_day since 1.5.0
+</p>
+<p>to_utc_timestamp since 1.5.0
+</p>
+<p>add_months since 1.5.0
+</p>
+<p>date_add since 1.5.0
+</p>
+<p>date_sub since 1.5.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other data time functions: <code><a href="column_datetime_functions.html">column_datetime_functions</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D dts &lt;- c(&quot;2005-01-02 18:47:22&quot;,
+##D         &quot;2005-12-24 16:30:58&quot;,
+##D         &quot;2005-10-28 07:30:05&quot;,
+##D         &quot;2005-12-28 07:01:05&quot;,
+##D         &quot;2006-01-24 00:01:10&quot;)
+##D y &lt;- c(2.0, 2.2, 3.4, 2.5, 1.8)
+##D df &lt;- createDataFrame(data.frame(time = as.POSIXct(dts), y = y))
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- createDataFrame(data.frame(time_string1 = as.POSIXct(dts),
+##D              time_string2 = as.POSIXct(dts[order(runif(length(dts)))])))
+##D tmp2 &lt;- mutate(tmp, datediff = datediff(tmp$time_string1, tmp$time_string2),
+##D                monthdiff = months_between(tmp$time_string1, tmp$time_string2))
+##D head(tmp2)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, from_utc = from_utc_timestamp(df$time, &quot;PST&quot;),
+##D                  to_utc = to_utc_timestamp(df$time, &quot;PST&quot;))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, t1 = add_months(df$time, 1),
+##D                   t2 = date_add(df$time, 2),
+##D                   t3 = date_sub(df$time, 3),
+##D                   t4 = next_day(df$time, &quot;Sun&quot;))
+##D head(tmp)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_datetime_functions.html b/site/docs/2.4.3/api/R/column_datetime_functions.html
new file mode 100644
index 0000000..2b46c92
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_datetime_functions.html
@@ -0,0 +1,403 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Date time functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_datetime_functions {SparkR}"><tr><td>column_datetime_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Date time functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Date time functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+current_date(x = "missing")
+
+current_timestamp(x = "missing")
+
+date_trunc(format, x)
+
+dayofmonth(x)
+
+dayofweek(x)
+
+dayofyear(x)
+
+from_unixtime(x, ...)
+
+hour(x)
+
+last_day(x)
+
+minute(x)
+
+month(x)
+
+quarter(x)
+
+second(x)
+
+to_date(x, format)
+
+to_timestamp(x, format)
+
+unix_timestamp(x, format)
+
+weekofyear(x)
+
+window(x, ...)
+
+year(x)
+
+## S4 method for signature 'Column'
+dayofmonth(x)
+
+## S4 method for signature 'Column'
+dayofweek(x)
+
+## S4 method for signature 'Column'
+dayofyear(x)
+
+## S4 method for signature 'Column'
+hour(x)
+
+## S4 method for signature 'Column'
+last_day(x)
+
+## S4 method for signature 'Column'
+minute(x)
+
+## S4 method for signature 'Column'
+month(x)
+
+## S4 method for signature 'Column'
+quarter(x)
+
+## S4 method for signature 'Column'
+second(x)
+
+## S4 method for signature 'Column,missing'
+to_date(x, format)
+
+## S4 method for signature 'Column,character'
+to_date(x, format)
+
+## S4 method for signature 'Column,missing'
+to_timestamp(x, format)
+
+## S4 method for signature 'Column,character'
+to_timestamp(x, format)
+
+## S4 method for signature 'Column'
+weekofyear(x)
+
+## S4 method for signature 'Column'
+year(x)
+
+## S4 method for signature 'Column'
+from_unixtime(x, format = "yyyy-MM-dd HH:mm:ss")
+
+## S4 method for signature 'Column'
+window(x, windowDuration, slideDuration = NULL,
+  startTime = NULL)
+
+## S4 method for signature 'missing,missing'
+unix_timestamp(x, format)
+
+## S4 method for signature 'Column,missing'
+unix_timestamp(x, format)
+
+## S4 method for signature 'Column,character'
+unix_timestamp(x,
+  format = "yyyy-MM-dd HH:mm:ss")
+
+## S4 method for signature 'Column'
+trunc(x, format)
+
+## S4 method for signature 'character,Column'
+date_trunc(format, x)
+
+## S4 method for signature 'missing'
+current_date()
+
+## S4 method for signature 'missing'
+current_timestamp()
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on. In <code>window</code>, it must be a time Column of
+<code>TimestampType</code>. This is not used with <code>current_date</code> and
+<code>current_timestamp</code></p>
+</td></tr>
+<tr valign="top"><td><code>format</code></td>
+<td>
+<p>The format for the given dates or timestamps in Column <code>x</code>. See the
+format used in the following methods:
+</p>
+
+<ul>
+<li> <p><code>to_date</code> and <code>to_timestamp</code>: it is the string to use to parse
+Column <code>x</code> to DateType or TimestampType.
+</p>
+</li>
+<li> <p><code>trunc</code>: it is the string to use to specify the truncation method.
+For example, &quot;year&quot;, &quot;yyyy&quot;, &quot;yy&quot; for truncate by year, or &quot;month&quot;, &quot;mon&quot;,
+&quot;mm&quot; for truncate by month.
+</p>
+</li>
+<li> <p><code>date_trunc</code>: it is similar with <code>trunc</code>'s but additionally
+supports &quot;day&quot;, &quot;dd&quot;, &quot;second&quot;, &quot;minute&quot;, &quot;hour&quot;, &quot;week&quot; and &quot;quarter&quot;.
+</p>
+</li></ul>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s).</p>
+</td></tr>
+<tr valign="top"><td><code>windowDuration</code></td>
+<td>
+<p>a string specifying the width of the window, e.g. '1 second',
+'1 day 12 hours', '2 minutes'. Valid interval strings are 'week',
+'day', 'hour', 'minute', 'second', 'millisecond', 'microsecond'. Note that
+the duration is a fixed length of time, and does not vary over time
+according to a calendar. For example, '1 day' always means 86,400,000
+milliseconds, not a calendar day.</p>
+</td></tr>
+<tr valign="top"><td><code>slideDuration</code></td>
+<td>
+<p>a string specifying the sliding interval of the window. Same format as
+<code>windowDuration</code>. A new window will be generated every
+<code>slideDuration</code>. Must be less than or equal to
+the <code>windowDuration</code>. This duration is likewise absolute, and does not
+vary according to a calendar.</p>
+</td></tr>
+<tr valign="top"><td><code>startTime</code></td>
+<td>
+<p>the offset with respect to 1970-01-01 00:00:00 UTC with which to start
+window intervals. For example, in order to have hourly tumbling windows
+that start 15 minutes past the hour, e.g. 12:15-13:15, 13:15-14:15... provide
+<code>startTime</code> as <code>"15 minutes"</code>.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>dayofmonth</code>: Extracts the day of the month as an integer from a
+given date/timestamp/string.
+</p>
+<p><code>dayofweek</code>: Extracts the day of the week as an integer from a
+given date/timestamp/string.
+</p>
+<p><code>dayofyear</code>: Extracts the day of the year as an integer from a
+given date/timestamp/string.
+</p>
+<p><code>hour</code>: Extracts the hour as an integer from a given date/timestamp/string.
+</p>
+<p><code>last_day</code>: Given a date column, returns the last day of the month which the
+given date belongs to. For example, input &quot;2015-07-27&quot; returns &quot;2015-07-31&quot; since
+July 31 is the last day of the month in July 2015.
+</p>
+<p><code>minute</code>: Extracts the minute as an integer from a given date/timestamp/string.
+</p>
+<p><code>month</code>: Extracts the month as an integer from a given date/timestamp/string.
+</p>
+<p><code>quarter</code>: Extracts the quarter as an integer from a given date/timestamp/string.
+</p>
+<p><code>second</code>: Extracts the second as an integer from a given date/timestamp/string.
+</p>
+<p><code>to_date</code>: Converts the column into a DateType. You may optionally specify
+a format according to the rules in:
+<a href="http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html">http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html</a>.
+If the string cannot be parsed according to the specified format (or default),
+the value of the column will be null.
+By default, it follows casting rules to a DateType if the format is omitted
+(equivalent to <code>cast(df$x, "date")</code>).
+</p>
+<p><code>to_timestamp</code>: Converts the column into a TimestampType. You may optionally specify
+a format according to the rules in:
+<a href="http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html">http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html</a>.
+If the string cannot be parsed according to the specified format (or default),
+the value of the column will be null.
+By default, it follows casting rules to a TimestampType if the format is omitted
+(equivalent to <code>cast(df$x, "timestamp")</code>).
+</p>
+<p><code>weekofyear</code>: Extracts the week number as an integer from a given date/timestamp/string.
+</p>
+<p><code>year</code>: Extracts the year as an integer from a given date/timestamp/string.
+</p>
+<p><code>from_unixtime</code>: Converts the number of seconds from unix epoch (1970-01-01 00:00:00 UTC)
+to a string representing the timestamp of that moment in the current system time zone in the JVM
+in the given format.
+See <a href="http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html">
+Customizing Formats</a> for available options.
+</p>
+<p><code>window</code>: Bucketizes rows into one or more time windows given a timestamp specifying column.
+Window starts are inclusive but the window ends are exclusive, e.g. 12:05 will be in the window
+[12:05,12:10) but not in [12:00,12:05). Windows can support microsecond precision. Windows in
+the order of months are not supported. It returns an output column of struct called 'window'
+by default with the nested columns 'start' and 'end'
+</p>
+<p><code>unix_timestamp</code>: Gets current Unix timestamp in seconds.
+</p>
+<p><code>trunc</code>: Returns date truncated to the unit specified by the format.
+</p>
+<p><code>date_trunc</code>: Returns timestamp truncated to the unit specified by the format.
+</p>
+<p><code>current_date</code>: Returns the current date as a date column.
+</p>
+<p><code>current_timestamp</code>: Returns the current timestamp as a timestamp column.
+</p>
+
+
+<h3>Note</h3>
+
+<p>dayofmonth since 1.5.0
+</p>
+<p>dayofweek since 2.3.0
+</p>
+<p>dayofyear since 1.5.0
+</p>
+<p>hour since 1.5.0
+</p>
+<p>last_day since 1.5.0
+</p>
+<p>minute since 1.5.0
+</p>
+<p>month since 1.5.0
+</p>
+<p>quarter since 1.5.0
+</p>
+<p>second since 1.5.0
+</p>
+<p>to_date(Column) since 1.5.0
+</p>
+<p>to_date(Column, character) since 2.2.0
+</p>
+<p>to_timestamp(Column) since 2.2.0
+</p>
+<p>to_timestamp(Column, character) since 2.2.0
+</p>
+<p>weekofyear since 1.5.0
+</p>
+<p>year since 1.5.0
+</p>
+<p>from_unixtime since 1.5.0
+</p>
+<p>window since 2.0.0
+</p>
+<p>unix_timestamp since 1.5.0
+</p>
+<p>unix_timestamp(Column) since 1.5.0
+</p>
+<p>unix_timestamp(Column, character) since 1.5.0
+</p>
+<p>trunc since 2.3.0
+</p>
+<p>date_trunc since 2.3.0
+</p>
+<p>current_date since 2.3.0
+</p>
+<p>current_timestamp since 2.3.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other data time functions: <code><a href="column_datetime_diff_functions.html">column_datetime_diff_functions</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D dts &lt;- c(&quot;2005-01-02 18:47:22&quot;,
+##D         &quot;2005-12-24 16:30:58&quot;,
+##D         &quot;2005-10-28 07:30:05&quot;,
+##D         &quot;2005-12-28 07:01:05&quot;,
+##D         &quot;2006-01-24 00:01:10&quot;)
+##D y &lt;- c(2.0, 2.2, 3.4, 2.5, 1.8)
+##D df &lt;- createDataFrame(data.frame(time = as.POSIXct(dts), y = y))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, df$time, year(df$time), quarter(df$time), month(df$time),
+##D             dayofmonth(df$time), dayofweek(df$time), dayofyear(df$time), weekofyear(df$time)))
+##D head(agg(groupBy(df, year(df$time)), count(df$y), avg(df$y)))
+##D head(agg(groupBy(df, month(df$time)), avg(df$y)))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, hour(df$time), minute(df$time), second(df$time)))
+##D head(agg(groupBy(df, dayofmonth(df$time)), avg(df$y)))
+##D head(agg(groupBy(df, hour(df$time)), avg(df$y)))
+##D head(agg(groupBy(df, minute(df$time)), avg(df$y)))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, df$time, last_day(df$time), month(df$time)))
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- createDataFrame(data.frame(time_string = dts))
+##D tmp2 &lt;- mutate(tmp, date1 = to_date(tmp$time_string),
+##D                    date2 = to_date(tmp$time_string, &quot;yyyy-MM-dd&quot;),
+##D                    date3 = date_format(tmp$time_string, &quot;MM/dd/yyy&quot;),
+##D                    time1 = to_timestamp(tmp$time_string),
+##D                    time2 = to_timestamp(tmp$time_string, &quot;yyyy-MM-dd&quot;))
+##D head(tmp2)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, to_unix = unix_timestamp(df$time),
+##D                   to_unix2 = unix_timestamp(df$time, &#39;yyyy-MM-dd HH&#39;),
+##D                   from_unix = from_unixtime(unix_timestamp(df$time)),
+##D                   from_unix2 = from_unixtime(unix_timestamp(df$time), &#39;yyyy-MM-dd HH:mm&#39;))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D # One minute windows every 15 seconds 10 seconds after the minute, e.g. 09:00:10-09:01:10,
+##D # 09:00:25-09:01:25, 09:00:40-09:01:40, ...
+##D window(df$time, &quot;1 minute&quot;, &quot;15 seconds&quot;, &quot;10 seconds&quot;)
+##D 
+##D # One minute tumbling windows 15 seconds after the minute, e.g. 09:00:15-09:01:15,
+##D # 09:01:15-09:02:15...
+##D window(df$time, &quot;1 minute&quot;, startTime = &quot;15 seconds&quot;)
+##D 
+##D # Thirty-second windows every 10 seconds, e.g. 09:00:00-09:00:30, 09:00:10-09:00:40, ...
+##D window(df$time, &quot;30 seconds&quot;, &quot;10 seconds&quot;)
+## End(Not run)
+
+## Not run: 
+##D head(select(df, df$time, trunc(df$time, &quot;year&quot;), trunc(df$time, &quot;yy&quot;),
+##D            trunc(df$time, &quot;month&quot;), trunc(df$time, &quot;mon&quot;)))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, df$time, date_trunc(&quot;hour&quot;, df$time), date_trunc(&quot;minute&quot;, df$time),
+##D             date_trunc(&quot;week&quot;, df$time), date_trunc(&quot;quarter&quot;, df$time)))
+## End(Not run)
+## Not run: 
+##D head(select(df, current_date(), current_timestamp()))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_math_functions.html b/site/docs/2.4.3/api/R/column_math_functions.html
new file mode 100644
index 0000000..ab383a3
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_math_functions.html
@@ -0,0 +1,415 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Math functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_math_functions {SparkR}"><tr><td>column_math_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Math functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Math functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+bin(x)
+
+bround(x, ...)
+
+cbrt(x)
+
+ceil(x)
+
+conv(x, fromBase, toBase)
+
+hex(x)
+
+hypot(y, x)
+
+pmod(y, x)
+
+rint(x)
+
+shiftLeft(y, x)
+
+shiftRight(y, x)
+
+shiftRightUnsigned(y, x)
+
+signum(x)
+
+toDegrees(x)
+
+toRadians(x)
+
+unhex(x)
+
+## S4 method for signature 'Column'
+abs(x)
+
+## S4 method for signature 'Column'
+acos(x)
+
+## S4 method for signature 'Column'
+asin(x)
+
+## S4 method for signature 'Column'
+atan(x)
+
+## S4 method for signature 'Column'
+bin(x)
+
+## S4 method for signature 'Column'
+cbrt(x)
+
+## S4 method for signature 'Column'
+ceil(x)
+
+## S4 method for signature 'Column'
+ceiling(x)
+
+## S4 method for signature 'Column'
+cos(x)
+
+## S4 method for signature 'Column'
+cosh(x)
+
+## S4 method for signature 'Column'
+exp(x)
+
+## S4 method for signature 'Column'
+expm1(x)
+
+## S4 method for signature 'Column'
+factorial(x)
+
+## S4 method for signature 'Column'
+floor(x)
+
+## S4 method for signature 'Column'
+hex(x)
+
+## S4 method for signature 'Column'
+log(x)
+
+## S4 method for signature 'Column'
+log10(x)
+
+## S4 method for signature 'Column'
+log1p(x)
+
+## S4 method for signature 'Column'
+log2(x)
+
+## S4 method for signature 'Column'
+rint(x)
+
+## S4 method for signature 'Column'
+round(x)
+
+## S4 method for signature 'Column'
+bround(x, scale = 0)
+
+## S4 method for signature 'Column'
+signum(x)
+
+## S4 method for signature 'Column'
+sign(x)
+
+## S4 method for signature 'Column'
+sin(x)
+
+## S4 method for signature 'Column'
+sinh(x)
+
+## S4 method for signature 'Column'
+sqrt(x)
+
+## S4 method for signature 'Column'
+tan(x)
+
+## S4 method for signature 'Column'
+tanh(x)
+
+## S4 method for signature 'Column'
+toDegrees(x)
+
+## S4 method for signature 'Column'
+toRadians(x)
+
+## S4 method for signature 'Column'
+unhex(x)
+
+## S4 method for signature 'Column'
+atan2(y, x)
+
+## S4 method for signature 'Column'
+hypot(y, x)
+
+## S4 method for signature 'Column'
+pmod(y, x)
+
+## S4 method for signature 'Column,numeric'
+shiftLeft(y, x)
+
+## S4 method for signature 'Column,numeric'
+shiftRight(y, x)
+
+## S4 method for signature 'Column,numeric'
+shiftRightUnsigned(y, x)
+
+## S4 method for signature 'Column,numeric,numeric'
+conv(x, fromBase, toBase)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on. In <code>shiftLeft</code>, <code>shiftRight</code> and
+<code>shiftRightUnsigned</code>, this is the number of bits to shift.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s).</p>
+</td></tr>
+<tr valign="top"><td><code>fromBase</code></td>
+<td>
+<p>base to convert from.</p>
+</td></tr>
+<tr valign="top"><td><code>toBase</code></td>
+<td>
+<p>base to convert to.</p>
+</td></tr>
+<tr valign="top"><td><code>y</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>scale</code></td>
+<td>
+<p>round to <code>scale</code> digits to the right of the decimal point when
+<code>scale</code> &gt; 0, the nearest even number when <code>scale</code> = 0, and <code>scale</code> digits
+to the left of the decimal point when <code>scale</code> &lt; 0.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>abs</code>: Computes the absolute value.
+</p>
+<p><code>acos</code>: Returns the inverse cosine of the given value,
+as if computed by <code>java.lang.Math.acos()</code>
+</p>
+<p><code>asin</code>: Returns the inverse sine of the given value,
+as if computed by <code>java.lang.Math.asin()</code>
+</p>
+<p><code>atan</code>: Returns the inverse tangent of the given value,
+as if computed by <code>java.lang.Math.atan()</code>
+</p>
+<p><code>bin</code>: Returns the string representation of the binary value
+of the given long column. For example, bin(&quot;12&quot;) returns &quot;1100&quot;.
+</p>
+<p><code>cbrt</code>: Computes the cube-root of the given value.
+</p>
+<p><code>ceil</code>: Computes the ceiling of the given value.
+</p>
+<p><code>ceiling</code>: Alias for <code>ceil</code>.
+</p>
+<p><code>cos</code>: Returns the cosine of the given value,
+as if computed by <code>java.lang.Math.cos()</code>. Units in radians.
+</p>
+<p><code>cosh</code>: Returns the hyperbolic cosine of the given value,
+as if computed by <code>java.lang.Math.cosh()</code>.
+</p>
+<p><code>exp</code>: Computes the exponential of the given value.
+</p>
+<p><code>expm1</code>: Computes the exponential of the given value minus one.
+</p>
+<p><code>factorial</code>: Computes the factorial of the given value.
+</p>
+<p><code>floor</code>: Computes the floor of the given value.
+</p>
+<p><code>hex</code>: Computes hex value of the given column.
+</p>
+<p><code>log</code>: Computes the natural logarithm of the given value.
+</p>
+<p><code>log10</code>: Computes the logarithm of the given value in base 10.
+</p>
+<p><code>log1p</code>: Computes the natural logarithm of the given value plus one.
+</p>
+<p><code>log2</code>: Computes the logarithm of the given column in base 2.
+</p>
+<p><code>rint</code>: Returns the double value that is closest in value to the argument and
+is equal to a mathematical integer.
+</p>
+<p><code>round</code>: Returns the value of the column rounded to 0 decimal places
+using HALF_UP rounding mode.
+</p>
+<p><code>bround</code>: Returns the value of the column <code>e</code> rounded to <code>scale</code> decimal places
+using HALF_EVEN rounding mode if <code>scale</code> &gt;= 0 or at integer part when <code>scale</code> &lt; 0.
+Also known as Gaussian rounding or bankers' rounding that rounds to the nearest even number.
+bround(2.5, 0) = 2, bround(3.5, 0) = 4.
+</p>
+<p><code>signum</code>: Computes the signum of the given value.
+</p>
+<p><code>sign</code>: Alias for <code>signum</code>.
+</p>
+<p><code>sin</code>: Returns the sine of the given value,
+as if computed by <code>java.lang.Math.sin()</code>. Units in radians.
+</p>
+<p><code>sinh</code>: Returns the hyperbolic sine of the given value,
+as if computed by <code>java.lang.Math.sinh()</code>.
+</p>
+<p><code>sqrt</code>: Computes the square root of the specified float value.
+</p>
+<p><code>tan</code>: Returns the tangent of the given value,
+as if computed by <code>java.lang.Math.tan()</code>.
+Units in radians.
+</p>
+<p><code>tanh</code>: Returns the hyperbolic tangent of the given value,
+as if computed by <code>java.lang.Math.tanh()</code>.
+</p>
+<p><code>toDegrees</code>: Converts an angle measured in radians to an approximately equivalent angle
+measured in degrees.
+</p>
+<p><code>toRadians</code>: Converts an angle measured in degrees to an approximately equivalent angle
+measured in radians.
+</p>
+<p><code>unhex</code>: Inverse of hex. Interprets each pair of characters as a hexadecimal number
+and converts to the byte representation of number.
+</p>
+<p><code>atan2</code>: Returns the angle theta from the conversion of rectangular coordinates
+(x, y) to polar coordinates (r, theta),
+as if computed by <code>java.lang.Math.atan2()</code>. Units in radians.
+</p>
+<p><code>hypot</code>: Computes &quot;sqrt(a^2 + b^2)&quot; without intermediate overflow or underflow.
+</p>
+<p><code>pmod</code>: Returns the positive value of dividend mod divisor.
+Column <code>x</code> is divisor column, and column <code>y</code> is the dividend column.
+</p>
+<p><code>shiftLeft</code>: Shifts the given value numBits left. If the given value is a long value,
+this function will return a long value else it will return an integer value.
+</p>
+<p><code>shiftRight</code>: (Signed) shifts the given value numBits right. If the given value is a long
+value, it will return a long value else it will return an integer value.
+</p>
+<p><code>shiftRightUnsigned</code>: (Unigned) shifts the given value numBits right. If the given value is
+a long value, it will return a long value else it will return an integer value.
+</p>
+<p><code>conv</code>: Converts a number in a string column from one base to another.
+</p>
+
+
+<h3>Note</h3>
+
+<p>abs since 1.5.0
+</p>
+<p>acos since 1.5.0
+</p>
+<p>asin since 1.5.0
+</p>
+<p>atan since 1.5.0
+</p>
+<p>bin since 1.5.0
+</p>
+<p>cbrt since 1.4.0
+</p>
+<p>ceil since 1.5.0
+</p>
+<p>ceiling since 1.5.0
+</p>
+<p>cos since 1.5.0
+</p>
+<p>cosh since 1.5.0
+</p>
+<p>exp since 1.5.0
+</p>
+<p>expm1 since 1.5.0
+</p>
+<p>factorial since 1.5.0
+</p>
+<p>floor since 1.5.0
+</p>
+<p>hex since 1.5.0
+</p>
+<p>log since 1.5.0
+</p>
+<p>log10 since 1.5.0
+</p>
+<p>log1p since 1.5.0
+</p>
+<p>log2 since 1.5.0
+</p>
+<p>rint since 1.5.0
+</p>
+<p>round since 1.5.0
+</p>
+<p>bround since 2.0.0
+</p>
+<p>signum since 1.5.0
+</p>
+<p>sign since 1.5.0
+</p>
+<p>sin since 1.5.0
+</p>
+<p>sinh since 1.5.0
+</p>
+<p>sqrt since 1.5.0
+</p>
+<p>tan since 1.5.0
+</p>
+<p>tanh since 1.5.0
+</p>
+<p>toDegrees since 1.4.0
+</p>
+<p>toRadians since 1.4.0
+</p>
+<p>unhex since 1.5.0
+</p>
+<p>atan2 since 1.5.0
+</p>
+<p>hypot since 1.4.0
+</p>
+<p>pmod since 1.5.0
+</p>
+<p>shiftLeft since 1.5.0
+</p>
+<p>shiftRight since 1.5.0
+</p>
+<p>shiftRightUnsigned since 1.5.0
+</p>
+<p>conv since 1.5.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+##D tmp &lt;- mutate(df, v1 = log(df$mpg), v2 = cbrt(df$disp),
+##D                   v3 = bround(df$wt, 1), v4 = bin(df$cyl),
+##D                   v5 = hex(df$wt), v6 = toDegrees(df$gear),
+##D                   v7 = atan2(df$cyl, df$am), v8 = hypot(df$cyl, df$am),
+##D                   v9 = pmod(df$hp, df$cyl), v10 = shiftLeft(df$disp, 1),
+##D                   v11 = conv(df$hp, 10, 16), v12 = sign(df$vs - 0.5),
+##D                   v13 = sqrt(df$disp), v14 = ceil(df$wt))
+##D head(tmp)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_misc_functions.html b/site/docs/2.4.3/api/R/column_misc_functions.html
new file mode 100644
index 0000000..658fda5
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_misc_functions.html
@@ -0,0 +1,117 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Miscellaneous functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_misc_functions {SparkR}"><tr><td>column_misc_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Miscellaneous functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Miscellaneous functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+crc32(x)
+
+hash(x, ...)
+
+md5(x)
+
+sha1(x)
+
+sha2(y, x)
+
+## S4 method for signature 'Column'
+crc32(x)
+
+## S4 method for signature 'Column'
+hash(x, ...)
+
+## S4 method for signature 'Column'
+md5(x)
+
+## S4 method for signature 'Column'
+sha1(x)
+
+## S4 method for signature 'Column,numeric'
+sha2(y, x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on. In <code>sha2</code>, it is one of 224, 256, 384, or 512.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional Columns.</p>
+</td></tr>
+<tr valign="top"><td><code>y</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>crc32</code>: Calculates the cyclic redundancy check value  (CRC32) of a binary column
+and returns the value as a bigint.
+</p>
+<p><code>hash</code>: Calculates the hash code of given columns, and returns the result
+as an int column.
+</p>
+<p><code>md5</code>: Calculates the MD5 digest of a binary column and returns the value
+as a 32 character hex string.
+</p>
+<p><code>sha1</code>: Calculates the SHA-1 digest of a binary column and returns the value
+as a 40 character hex string.
+</p>
+<p><code>sha2</code>: Calculates the SHA-2 family of hash functions of a binary column and
+returns the value as a hex string. The second argument <code>x</code> specifies the number
+of bits, and is one of 224, 256, 384, or 512.
+</p>
+
+
+<h3>Note</h3>
+
+<p>crc32 since 1.5.0
+</p>
+<p>hash since 2.0.0
+</p>
+<p>md5 since 1.5.0
+</p>
+<p>sha1 since 1.5.0
+</p>
+<p>sha2 since 1.5.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars)[, 1:2])
+##D tmp &lt;- mutate(df, v1 = crc32(df$model), v2 = hash(df$model),
+##D                   v3 = hash(df$model, df$mpg), v4 = md5(df$model),
+##D                   v5 = sha1(df$model), v6 = sha2(df$model, 256))
+##D head(tmp)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_nonaggregate_functions.html b/site/docs/2.4.3/api/R/column_nonaggregate_functions.html
new file mode 100644
index 0000000..e663e4d
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_nonaggregate_functions.html
@@ -0,0 +1,351 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Non-aggregate functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_nonaggregate_functions {SparkR}"><tr><td>column_nonaggregate_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Non-aggregate functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Non-aggregate functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+when(condition, value)
+
+bitwiseNOT(x)
+
+create_array(x, ...)
+
+create_map(x, ...)
+
+expr(x)
+
+greatest(x, ...)
+
+input_file_name(x = "missing")
+
+isnan(x)
+
+least(x, ...)
+
+lit(x)
+
+monotonically_increasing_id(x = "missing")
+
+nanvl(y, x)
+
+negate(x)
+
+rand(seed)
+
+randn(seed)
+
+spark_partition_id(x = "missing")
+
+struct(x, ...)
+
+## S4 method for signature 'ANY'
+lit(x)
+
+## S4 method for signature 'Column'
+bitwiseNOT(x)
+
+## S4 method for signature 'Column'
+coalesce(x, ...)
+
+## S4 method for signature 'Column'
+isnan(x)
+
+## S4 method for signature 'Column'
+is.nan(x)
+
+## S4 method for signature 'missing'
+monotonically_increasing_id()
+
+## S4 method for signature 'Column'
+negate(x)
+
+## S4 method for signature 'missing'
+spark_partition_id()
+
+## S4 method for signature 'characterOrColumn'
+struct(x, ...)
+
+## S4 method for signature 'Column'
+nanvl(y, x)
+
+## S4 method for signature 'Column'
+greatest(x, ...)
+
+## S4 method for signature 'Column'
+least(x, ...)
+
+## S4 method for signature 'character'
+expr(x)
+
+## S4 method for signature 'missing'
+rand(seed)
+
+## S4 method for signature 'numeric'
+rand(seed)
+
+## S4 method for signature 'missing'
+randn(seed)
+
+## S4 method for signature 'numeric'
+randn(seed)
+
+## S4 method for signature 'Column'
+when(condition, value)
+
+## S4 method for signature 'Column'
+ifelse(test, yes, no)
+
+## S4 method for signature 'Column'
+create_array(x, ...)
+
+## S4 method for signature 'Column'
+create_map(x, ...)
+
+## S4 method for signature 'missing'
+input_file_name()
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>condition</code></td>
+<td>
+<p>the condition to test on. Must be a Column expression.</p>
+</td></tr>
+<tr valign="top"><td><code>value</code></td>
+<td>
+<p>result expression.</p>
+</td></tr>
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on. In <code>lit</code>, it is a literal value or a Column.
+In <code>expr</code>, it contains an expression character object to be parsed.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional Columns.</p>
+</td></tr>
+<tr valign="top"><td><code>y</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>seed</code></td>
+<td>
+<p>a random seed. Can be missing.</p>
+</td></tr>
+<tr valign="top"><td><code>test</code></td>
+<td>
+<p>a Column expression that describes the condition.</p>
+</td></tr>
+<tr valign="top"><td><code>yes</code></td>
+<td>
+<p>return values for <code>TRUE</code> elements of test.</p>
+</td></tr>
+<tr valign="top"><td><code>no</code></td>
+<td>
+<p>return values for <code>FALSE</code> elements of test.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>lit</code>: A new Column is created to represent the literal value.
+If the parameter is a Column, it is returned unchanged.
+</p>
+<p><code>bitwiseNOT</code>: Computes bitwise NOT.
+</p>
+<p><code>coalesce</code>: Returns the first column that is not NA, or NA if all inputs are.
+</p>
+<p><code>isnan</code>: Returns true if the column is NaN.
+</p>
+<p><code>is.nan</code>: Alias for <a href="column_nonaggregate_functions.html">isnan</a>.
+</p>
+<p><code>monotonically_increasing_id</code>: Returns a column that generates monotonically increasing
+64-bit integers. The generated ID is guaranteed to be monotonically increasing and unique,
+but not consecutive. The current implementation puts the partition ID in the upper 31 bits,
+and the record number within each partition in the lower 33 bits. The assumption is that the
+SparkDataFrame has less than 1 billion partitions, and each partition has less than 8 billion
+records. As an example, consider a SparkDataFrame with two partitions, each with 3 records.
+This expression would return the following IDs:
+0, 1, 2, 8589934592 (1L &lt;&lt; 33), 8589934593, 8589934594.
+This is equivalent to the MONOTONICALLY_INCREASING_ID function in SQL.
+The method should be used with no argument.
+Note: the function is non-deterministic because its result depends on partition IDs.
+</p>
+<p><code>negate</code>: Unary minus, i.e. negate the expression.
+</p>
+<p><code>spark_partition_id</code>: Returns the partition ID as a SparkDataFrame column.
+Note that this is nondeterministic because it depends on data partitioning and
+task scheduling.
+This is equivalent to the <code>SPARK_PARTITION_ID</code> function in SQL.
+</p>
+<p><code>struct</code>: Creates a new struct column that composes multiple input columns.
+</p>
+<p><code>nanvl</code>: Returns the first column (<code>y</code>) if it is not NaN, or the second column
+(<code>x</code>) if the first column is NaN. Both inputs should be floating point columns
+(DoubleType or FloatType).
+</p>
+<p><code>greatest</code>: Returns the greatest value of the list of column names, skipping null values.
+This function takes at least 2 parameters. It will return null if all parameters are null.
+</p>
+<p><code>least</code>: Returns the least value of the list of column names, skipping null values.
+This function takes at least 2 parameters. It will return null if all parameters are null.
+</p>
+<p><code>expr</code>: Parses the expression string into the column that it represents, similar to
+<code>SparkDataFrame.selectExpr</code>
+</p>
+<p><code>rand</code>: Generates a random column with independent and identically distributed (i.i.d.)
+samples from U[0.0, 1.0].
+Note: the function is non-deterministic in general case.
+</p>
+<p><code>randn</code>: Generates a column with independent and identically distributed (i.i.d.) samples
+from the standard normal distribution.
+Note: the function is non-deterministic in general case.
+</p>
+<p><code>when</code>: Evaluates a list of conditions and returns one of multiple possible result
+expressions. For unmatched expressions null is returned.
+</p>
+<p><code>ifelse</code>: Evaluates a list of conditions and returns <code>yes</code> if the conditions are
+satisfied. Otherwise <code>no</code> is returned for unmatched conditions.
+</p>
+<p><code>create_array</code>: Creates a new array column. The input columns must all have the same data
+type.
+</p>
+<p><code>create_map</code>: Creates a new map column. The input columns must be grouped as key-value
+pairs, e.g. (key1, value1, key2, value2, ...).
+The key columns must all have the same data type, and can't be null.
+The value columns must all have the same data type.
+</p>
+<p><code>input_file_name</code>: Creates a string column with the input file name for a given row.
+The method should be used with no argument.
+</p>
+
+
+<h3>Note</h3>
+
+<p>lit since 1.5.0
+</p>
+<p>bitwiseNOT since 1.5.0
+</p>
+<p>coalesce(Column) since 2.1.1
+</p>
+<p>isnan since 2.0.0
+</p>
+<p>is.nan since 2.0.0
+</p>
+<p>negate since 1.5.0
+</p>
+<p>spark_partition_id since 2.0.0
+</p>
+<p>struct since 1.6.0
+</p>
+<p>nanvl since 1.5.0
+</p>
+<p>greatest since 1.5.0
+</p>
+<p>least since 1.5.0
+</p>
+<p>expr since 1.5.0
+</p>
+<p>rand since 1.5.0
+</p>
+<p>rand(numeric) since 1.5.0
+</p>
+<p>randn since 1.5.0
+</p>
+<p>randn(numeric) since 1.5.0
+</p>
+<p>when since 1.5.0
+</p>
+<p>ifelse since 1.5.0
+</p>
+<p>create_array since 2.3.0
+</p>
+<p>create_map since 2.3.0
+</p>
+<p>input_file_name since 2.3.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>coalesce,SparkDataFrame-method
+</p>
+<p>Other non-aggregate functions: <code><a href="column.html">column</a></code>,
+<code><a href="not.html">not</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, v1 = lit(df$mpg), v2 = lit(&quot;x&quot;), v3 = lit(&quot;2015-01-01&quot;),
+##D                   v4 = negate(df$mpg), v5 = expr(&#39;length(model)&#39;),
+##D                   v6 = greatest(df$vs, df$am), v7 = least(df$vs, df$am),
+##D                   v8 = column(&quot;mpg&quot;))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D head(select(df, bitwiseNOT(cast(df$vs, &quot;int&quot;))))
+## End(Not run)
+
+## Not run: head(select(df, monotonically_increasing_id()))
+
+## Not run: head(select(df, spark_partition_id()))
+
+## Not run: 
+##D tmp &lt;- mutate(df, v1 = struct(df$mpg, df$cyl), v2 = struct(&quot;hp&quot;, &quot;wt&quot;, &quot;vs&quot;),
+##D                   v3 = create_array(df$mpg, df$cyl, df$hp),
+##D                   v4 = create_map(lit(&quot;x&quot;), lit(1.0), lit(&quot;y&quot;), lit(-1.0)))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, r1 = rand(), r2 = rand(10), r3 = randn(), r4 = randn(10))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, mpg_na = otherwise(when(df$mpg &gt; 20, df$mpg), lit(NaN)),
+##D                   mpg2 = ifelse(df$mpg &gt; 20 &amp; df$am &gt; 0, 0, 1),
+##D                   mpg3 = ifelse(df$mpg &gt; 20, df$mpg, 20.0))
+##D head(tmp)
+##D tmp &lt;- mutate(tmp, ind_na1 = is.nan(tmp$mpg_na), ind_na2 = isnan(tmp$mpg_na))
+##D head(select(tmp, coalesce(tmp$mpg_na, tmp$mpg)))
+##D head(select(tmp, nanvl(tmp$mpg_na, tmp$hp)))
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- read.text(&quot;README.md&quot;)
+##D head(select(tmp, input_file_name()))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_string_functions.html b/site/docs/2.4.3/api/R/column_string_functions.html
new file mode 100644
index 0000000..3997b0a
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_string_functions.html
@@ -0,0 +1,522 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: String functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_string_functions {SparkR}"><tr><td>column_string_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>String functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>String functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+ascii(x)
+
+base64(x)
+
+concat_ws(sep, x, ...)
+
+decode(x, charset)
+
+encode(x, charset)
+
+format_number(y, x)
+
+format_string(format, x, ...)
+
+initcap(x)
+
+instr(y, x)
+
+levenshtein(y, x)
+
+locate(substr, str, ...)
+
+lower(x)
+
+lpad(x, len, pad)
+
+ltrim(x, trimString)
+
+regexp_extract(x, pattern, idx)
+
+regexp_replace(x, pattern, replacement)
+
+repeat_string(x, n)
+
+rpad(x, len, pad)
+
+rtrim(x, trimString)
+
+split_string(x, pattern)
+
+soundex(x)
+
+substring_index(x, delim, count)
+
+translate(x, matchingString, replaceString)
+
+trim(x, trimString)
+
+unbase64(x)
+
+upper(x)
+
+## S4 method for signature 'Column'
+ascii(x)
+
+## S4 method for signature 'Column'
+base64(x)
+
+## S4 method for signature 'Column,character'
+decode(x, charset)
+
+## S4 method for signature 'Column,character'
+encode(x, charset)
+
+## S4 method for signature 'Column'
+initcap(x)
+
+## S4 method for signature 'Column'
+length(x)
+
+## S4 method for signature 'Column'
+lower(x)
+
+## S4 method for signature 'Column,missing'
+ltrim(x, trimString)
+
+## S4 method for signature 'Column,character'
+ltrim(x, trimString)
+
+## S4 method for signature 'Column,missing'
+rtrim(x, trimString)
+
+## S4 method for signature 'Column,character'
+rtrim(x, trimString)
+
+## S4 method for signature 'Column'
+soundex(x)
+
+## S4 method for signature 'Column,missing'
+trim(x, trimString)
+
+## S4 method for signature 'Column,character'
+trim(x, trimString)
+
+## S4 method for signature 'Column'
+unbase64(x)
+
+## S4 method for signature 'Column'
+upper(x)
+
+## S4 method for signature 'Column'
+levenshtein(y, x)
+
+## S4 method for signature 'Column,character'
+instr(y, x)
+
+## S4 method for signature 'Column,numeric'
+format_number(y, x)
+
+## S4 method for signature 'character,Column'
+concat_ws(sep, x, ...)
+
+## S4 method for signature 'character,Column'
+format_string(format, x, ...)
+
+## S4 method for signature 'character,Column'
+locate(substr, str, pos = 1)
+
+## S4 method for signature 'Column,numeric,character'
+lpad(x, len, pad)
+
+## S4 method for signature 'Column,character,numeric'
+regexp_extract(x, pattern, idx)
+
+## S4 method for signature 'Column,character,character'
+regexp_replace(x, pattern,
+  replacement)
+
+## S4 method for signature 'Column,numeric,character'
+rpad(x, len, pad)
+
+## S4 method for signature 'Column,character,numeric'
+substring_index(x, delim, count)
+
+## S4 method for signature 'Column,character,character'
+translate(x, matchingString,
+  replaceString)
+
+## S4 method for signature 'Column,character'
+split_string(x, pattern)
+
+## S4 method for signature 'Column,numeric'
+repeat_string(x, n)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>Column to compute on except in the following methods:
+</p>
+
+<ul>
+<li> <p><code>instr</code>: <code>character</code>, the substring to check. See 'Details'.
+</p>
+</li>
+<li> <p><code>format_number</code>: <code>numeric</code>, the number of decimal place to
+format to. See 'Details'.
+</p>
+</li></ul>
+</td></tr>
+<tr valign="top"><td><code>sep</code></td>
+<td>
+<p>separator to use.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional Columns.</p>
+</td></tr>
+<tr valign="top"><td><code>charset</code></td>
+<td>
+<p>character set to use (one of &quot;US-ASCII&quot;, &quot;ISO-8859-1&quot;, &quot;UTF-8&quot;, &quot;UTF-16BE&quot;,
+&quot;UTF-16LE&quot;, &quot;UTF-16&quot;).</p>
+</td></tr>
+<tr valign="top"><td><code>y</code></td>
+<td>
+<p>Column to compute on.</p>
+</td></tr>
+<tr valign="top"><td><code>format</code></td>
+<td>
+<p>a character object of format strings.</p>
+</td></tr>
+<tr valign="top"><td><code>substr</code></td>
+<td>
+<p>a character string to be matched.</p>
+</td></tr>
+<tr valign="top"><td><code>str</code></td>
+<td>
+<p>a Column where matches are sought for each entry.</p>
+</td></tr>
+<tr valign="top"><td><code>len</code></td>
+<td>
+<p>maximum length of each output result.</p>
+</td></tr>
+<tr valign="top"><td><code>pad</code></td>
+<td>
+<p>a character string to be padded with.</p>
+</td></tr>
+<tr valign="top"><td><code>trimString</code></td>
+<td>
+<p>a character string to trim with</p>
+</td></tr>
+<tr valign="top"><td><code>pattern</code></td>
+<td>
+<p>a regular expression.</p>
+</td></tr>
+<tr valign="top"><td><code>idx</code></td>
+<td>
+<p>a group index.</p>
+</td></tr>
+<tr valign="top"><td><code>replacement</code></td>
+<td>
+<p>a character string that a matched <code>pattern</code> is replaced with.</p>
+</td></tr>
+<tr valign="top"><td><code>n</code></td>
+<td>
+<p>number of repetitions.</p>
+</td></tr>
+<tr valign="top"><td><code>delim</code></td>
+<td>
+<p>a delimiter string.</p>
+</td></tr>
+<tr valign="top"><td><code>count</code></td>
+<td>
+<p>number of occurrences of <code>delim</code> before the substring is returned.
+A positive number means counting from the left, while negative means
+counting from the right.</p>
+</td></tr>
+<tr valign="top"><td><code>matchingString</code></td>
+<td>
+<p>a source string where each character will be translated.</p>
+</td></tr>
+<tr valign="top"><td><code>replaceString</code></td>
+<td>
+<p>a target string where each <code>matchingString</code> character will
+be replaced by the character in <code>replaceString</code>
+at the same location, if any.</p>
+</td></tr>
+<tr valign="top"><td><code>pos</code></td>
+<td>
+<p>start position of search.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>ascii</code>: Computes the numeric value of the first character of the string column,
+and returns the result as an int column.
+</p>
+<p><code>base64</code>: Computes the BASE64 encoding of a binary column and returns it as
+a string column. This is the reverse of unbase64.
+</p>
+<p><code>decode</code>: Computes the first argument into a string from a binary using the provided
+character set.
+</p>
+<p><code>encode</code>: Computes the first argument into a binary from a string using the provided
+character set.
+</p>
+<p><code>initcap</code>: Returns a new string column by converting the first letter of
+each word to uppercase. Words are delimited by whitespace. For example, &quot;hello world&quot;
+will become &quot;Hello World&quot;.
+</p>
+<p><code>length</code>: Computes the character length of a string data or number of bytes
+of a binary data. The length of string data includes the trailing spaces.
+The length of binary data includes binary zeros.
+</p>
+<p><code>lower</code>: Converts a string column to lower case.
+</p>
+<p><code>ltrim</code>: Trims the spaces from left end for the specified string value. Optionally a
+<code>trimString</code> can be specified.
+</p>
+<p><code>rtrim</code>: Trims the spaces from right end for the specified string value. Optionally a
+<code>trimString</code> can be specified.
+</p>
+<p><code>soundex</code>: Returns the soundex code for the specified expression.
+</p>
+<p><code>trim</code>: Trims the spaces from both ends for the specified string column. Optionally a
+<code>trimString</code> can be specified.
+</p>
+<p><code>unbase64</code>: Decodes a BASE64 encoded string column and returns it as a binary column.
+This is the reverse of base64.
+</p>
+<p><code>upper</code>: Converts a string column to upper case.
+</p>
+<p><code>levenshtein</code>: Computes the Levenshtein distance of the two given string columns.
+</p>
+<p><code>instr</code>: Locates the position of the first occurrence of a substring (<code>x</code>)
+in the given string column (<code>y</code>). Returns null if either of the arguments are null.
+Note: The position is not zero based, but 1 based index. Returns 0 if the substring
+could not be found in the string column.
+</p>
+<p><code>format_number</code>: Formats numeric column <code>y</code> to a format like '#,###,###.##',
+rounded to <code>x</code> decimal places with HALF_EVEN round mode, and returns the result
+as a string column.
+If <code>x</code> is 0, the result has no decimal point or fractional part.
+If <code>x</code> &lt; 0, the result will be null.
+</p>
+<p><code>concat_ws</code>: Concatenates multiple input string columns together into a single
+string column, using the given separator.
+</p>
+<p><code>format_string</code>: Formats the arguments in printf-style and returns the result
+as a string column.
+</p>
+<p><code>locate</code>: Locates the position of the first occurrence of substr.
+Note: The position is not zero based, but 1 based index. Returns 0 if substr
+could not be found in str.
+</p>
+<p><code>lpad</code>: Left-padded with pad to a length of len.
+</p>
+<p><code>regexp_extract</code>: Extracts a specific <code>idx</code> group identified by a Java regex,
+from the specified string column. If the regex did not match, or the specified group did
+not match, an empty string is returned.
+</p>
+<p><code>regexp_replace</code>: Replaces all substrings of the specified string value that
+match regexp with rep.
+</p>
+<p><code>rpad</code>: Right-padded with pad to a length of len.
+</p>
+<p><code>substring_index</code>: Returns the substring from string (<code>x</code>) before <code>count</code>
+occurrences of the delimiter (<code>delim</code>). If <code>count</code> is positive, everything the left of
+the final delimiter (counting from left) is returned. If <code>count</code> is negative, every to the
+right of the final delimiter (counting from the right) is returned. <code>substring_index</code>
+performs a case-sensitive match when searching for the delimiter.
+</p>
+<p><code>translate</code>: Translates any character in the src by a character in replaceString.
+The characters in replaceString is corresponding to the characters in matchingString.
+The translate will happen when any character in the string matching with the character
+in the matchingString.
+</p>
+<p><code>split_string</code>: Splits string on regular expression.
+Equivalent to <code>split</code> SQL function.
+</p>
+<p><code>repeat_string</code>: Repeats string n times.
+Equivalent to <code>repeat</code> SQL function.
+</p>
+
+
+<h3>Note</h3>
+
+<p>ascii since 1.5.0
+</p>
+<p>base64 since 1.5.0
+</p>
+<p>decode since 1.6.0
+</p>
+<p>encode since 1.6.0
+</p>
+<p>initcap since 1.5.0
+</p>
+<p>length since 1.5.0
+</p>
+<p>lower since 1.4.0
+</p>
+<p>ltrim since 1.5.0
+</p>
+<p>ltrim(Column, character) since 2.3.0
+</p>
+<p>rtrim since 1.5.0
+</p>
+<p>rtrim(Column, character) since 2.3.0
+</p>
+<p>soundex since 1.5.0
+</p>
+<p>trim since 1.5.0
+</p>
+<p>trim(Column, character) since 2.3.0
+</p>
+<p>unbase64 since 1.5.0
+</p>
+<p>upper since 1.4.0
+</p>
+<p>levenshtein since 1.5.0
+</p>
+<p>instr since 1.5.0
+</p>
+<p>format_number since 1.5.0
+</p>
+<p>concat_ws since 1.5.0
+</p>
+<p>format_string since 1.5.0
+</p>
+<p>locate since 1.5.0
+</p>
+<p>lpad since 1.5.0
+</p>
+<p>regexp_extract since 1.5.0
+</p>
+<p>regexp_replace since 1.5.0
+</p>
+<p>rpad since 1.5.0
+</p>
+<p>substring_index since 1.5.0
+</p>
+<p>translate since 1.5.0
+</p>
+<p>split_string 2.3.0
+</p>
+<p>repeat_string since 2.3.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(as.data.frame(Titanic, stringsAsFactors = FALSE))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, ascii(df$Class), ascii(df$Sex)))
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, s1 = encode(df$Class, &quot;UTF-8&quot;))
+##D str(tmp)
+##D tmp2 &lt;- mutate(tmp, s2 = base64(tmp$s1), s3 = decode(tmp$s1, &quot;UTF-8&quot;),
+##D                     s4 = soundex(tmp$Sex))
+##D head(tmp2)
+##D head(select(tmp2, unbase64(tmp2$s2)))
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, sex_lower = lower(df$Sex), age_upper = upper(df$age),
+##D                   sex_age = concat_ws(&quot; &quot;, lower(df$sex), lower(df$age)))
+##D head(tmp)
+##D tmp2 &lt;- mutate(tmp, s1 = initcap(tmp$sex_lower), s2 = initcap(tmp$sex_age),
+##D                     s3 = reverse(df$Sex))
+##D head(tmp2)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, SexLpad = lpad(df$Sex, 6, &quot; &quot;), SexRpad = rpad(df$Sex, 7, &quot; &quot;))
+##D head(select(tmp, length(tmp$Sex), length(tmp$SexLpad), length(tmp$SexRpad)))
+##D tmp2 &lt;- mutate(tmp, SexLtrim = ltrim(tmp$SexLpad), SexRtrim = rtrim(tmp$SexRpad),
+##D                     SexTrim = trim(tmp$SexLpad))
+##D head(select(tmp2, length(tmp2$Sex), length(tmp2$SexLtrim),
+##D                   length(tmp2$SexRtrim), length(tmp2$SexTrim)))
+##D 
+##D tmp &lt;- mutate(df, SexLpad = lpad(df$Sex, 6, &quot;xx&quot;), SexRpad = rpad(df$Sex, 7, &quot;xx&quot;))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, d1 = levenshtein(df$Class, df$Sex),
+##D                   d2 = levenshtein(df$Age, df$Sex),
+##D                   d3 = levenshtein(df$Age, df$Age))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, s1 = instr(df$Sex, &quot;m&quot;), s2 = instr(df$Sex, &quot;M&quot;),
+##D                   s3 = locate(&quot;m&quot;, df$Sex), s4 = locate(&quot;m&quot;, df$Sex, pos = 4))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, v1 = df$Freq/3)
+##D head(select(tmp, format_number(tmp$v1, 0), format_number(tmp$v1, 2),
+##D                  format_string(&quot;%4.2f %s&quot;, tmp$v1, tmp$Sex)), 10)
+## End(Not run)
+
+## Not run: 
+##D # concatenate strings
+##D tmp &lt;- mutate(df, s1 = concat_ws(&quot;_&quot;, df$Class, df$Sex),
+##D                   s2 = concat_ws(&quot;+&quot;, df$Class, df$Sex, df$Age, df$Survived))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D tmp &lt;- mutate(df, s1 = regexp_extract(df$Class, &quot;(\\d+)\\w+&quot;, 1),
+##D                   s2 = regexp_extract(df$Sex, &quot;^(\\w)\\w+&quot;, 1),
+##D                   s3 = regexp_replace(df$Class, &quot;\\D+&quot;, &quot;&quot;),
+##D                   s4 = substring_index(df$Sex, &quot;a&quot;, 1),
+##D                   s5 = substring_index(df$Sex, &quot;a&quot;, -1),
+##D                   s6 = translate(df$Sex, &quot;ale&quot;, &quot;&quot;),
+##D                   s7 = translate(df$Sex, &quot;a&quot;, &quot;-&quot;))
+##D head(tmp)
+## End(Not run)
+
+## Not run: 
+##D head(select(df, split_string(df$Sex, &quot;a&quot;)))
+##D head(select(df, split_string(df$Class, &quot;\\d&quot;)))
+##D # This is equivalent to the following SQL expression
+##D head(selectExpr(df, &quot;split(Class, &#39;\\\\d&#39;)&quot;))
+## End(Not run)
+
+## Not run: 
+##D head(select(df, repeat_string(df$Class, 3)))
+##D # This is equivalent to the following SQL expression
+##D head(selectExpr(df, &quot;repeat(Class, 3)&quot;))
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/column_window_functions.html b/site/docs/2.4.3/api/R/column_window_functions.html
new file mode 100644
index 0000000..25096de
--- /dev/null
+++ b/site/docs/2.4.3/api/R/column_window_functions.html
@@ -0,0 +1,187 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Window functions for Column operations</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for column_window_functions {SparkR}"><tr><td>column_window_functions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Window functions for Column operations</h2>
+
+<h3>Description</h3>
+
+<p>Window functions defined for <code>Column</code>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+cume_dist(x = "missing")
+
+dense_rank(x = "missing")
+
+lag(x, ...)
+
+lead(x, offset, defaultValue = NULL)
+
+ntile(x)
+
+percent_rank(x = "missing")
+
+rank(x, ...)
+
+row_number(x = "missing")
+
+## S4 method for signature 'missing'
+cume_dist()
+
+## S4 method for signature 'missing'
+dense_rank()
+
+## S4 method for signature 'characterOrColumn'
+lag(x, offset = 1, defaultValue = NULL)
+
+## S4 method for signature 'characterOrColumn,numeric'
+lead(x, offset = 1,
+  defaultValue = NULL)
+
+## S4 method for signature 'numeric'
+ntile(x)
+
+## S4 method for signature 'missing'
+percent_rank()
+
+## S4 method for signature 'missing'
+rank()
+
+## S4 method for signature 'ANY'
+rank(x, ...)
+
+## S4 method for signature 'missing'
+row_number()
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>In <code>lag</code> and <code>lead</code>, it is the column as a character string or a Column
+to compute on. In <code>ntile</code>, it is the number of ntile groups.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s).</p>
+</td></tr>
+<tr valign="top"><td><code>offset</code></td>
+<td>
+<p>In <code>lag</code>, the number of rows back from the current row from which to obtain
+a value. In <code>lead</code>, the number of rows after the current row from which to
+obtain a value. If not specified, the default is 1.</p>
+</td></tr>
+<tr valign="top"><td><code>defaultValue</code></td>
+<td>
+<p>(optional) default to use when the offset row does not exist.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>cume_dist</code>: Returns the cumulative distribution of values within a window partition,
+i.e. the fraction of rows that are below the current row:
+(number of values before and including x) / (total number of rows in the partition).
+This is equivalent to the <code>CUME_DIST</code> function in SQL.
+The method should be used with no argument.
+</p>
+<p><code>dense_rank</code>: Returns the rank of rows within a window partition, without any gaps.
+The difference between rank and dense_rank is that dense_rank leaves no gaps in ranking
+sequence when there are ties. That is, if you were ranking a competition using dense_rank
+and had three people tie for second place, you would say that all three were in second
+place and that the next person came in third. Rank would give me sequential numbers, making
+the person that came in third place (after the ties) would register as coming in fifth.
+This is equivalent to the <code>DENSE_RANK</code> function in SQL.
+The method should be used with no argument.
+</p>
+<p><code>lag</code>: Returns the value that is <code>offset</code> rows before the current row, and
+<code>defaultValue</code> if there is less than <code>offset</code> rows before the current row. For example,
+an <code>offset</code> of one will return the previous row at any given point in the window partition.
+This is equivalent to the <code>LAG</code> function in SQL.
+</p>
+<p><code>lead</code>: Returns the value that is <code>offset</code> rows after the current row, and
+<code>defaultValue</code> if there is less than <code>offset</code> rows after the current row.
+For example, an <code>offset</code> of one will return the next row at any given point
+in the window partition.
+This is equivalent to the <code>LEAD</code> function in SQL.
+</p>
+<p><code>ntile</code>: Returns the ntile group id (from 1 to n inclusive) in an ordered window
+partition. For example, if n is 4, the first quarter of the rows will get value 1, the second
+quarter will get 2, the third quarter will get 3, and the last quarter will get 4.
+This is equivalent to the <code>NTILE</code> function in SQL.
+</p>
+<p><code>percent_rank</code>: Returns the relative rank (i.e. percentile) of rows within a window
+partition.
+This is computed by: (rank of row in its partition - 1) / (number of rows in the partition - 1).
+This is equivalent to the <code>PERCENT_RANK</code> function in SQL.
+The method should be used with no argument.
+</p>
+<p><code>rank</code>: Returns the rank of rows within a window partition.
+The difference between rank and dense_rank is that dense_rank leaves no gaps in ranking
+sequence when there are ties. That is, if you were ranking a competition using dense_rank
+and had three people tie for second place, you would say that all three were in second
+place and that the next person came in third. Rank would give me sequential numbers, making
+the person that came in third place (after the ties) would register as coming in fifth.
+This is equivalent to the <code>RANK</code> function in SQL.
+The method should be used with no argument.
+</p>
+<p><code>row_number</code>: Returns a sequential number starting at 1 within a window partition.
+This is equivalent to the <code>ROW_NUMBER</code> function in SQL.
+The method should be used with no argument.
+</p>
+
+
+<h3>Note</h3>
+
+<p>cume_dist since 1.6.0
+</p>
+<p>dense_rank since 1.6.0
+</p>
+<p>lag since 1.6.0
+</p>
+<p>lead since 1.6.0
+</p>
+<p>ntile since 1.6.0
+</p>
+<p>percent_rank since 1.6.0
+</p>
+<p>rank since 1.6.0
+</p>
+<p>row_number since 1.6.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D # Dataframe used throughout this doc
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+##D ws &lt;- orderBy(windowPartitionBy(&quot;am&quot;), &quot;hp&quot;)
+##D tmp &lt;- mutate(df, dist = over(cume_dist(), ws), dense_rank = over(dense_rank(), ws),
+##D               lag = over(lag(df$mpg), ws), lead = over(lead(df$mpg, 1), ws),
+##D               percent_rank = over(percent_rank(), ws),
+##D               rank = over(rank(), ws), row_number = over(row_number(), ws))
+##D # Get ntile group id (1-4) for hp
+##D tmp &lt;- mutate(tmp, ntile = over(ntile(4), ws))
+##D head(tmp)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/columnfunctions.html b/site/docs/2.4.3/api/R/columnfunctions.html
new file mode 100644
index 0000000..b9ff0f0
--- /dev/null
+++ b/site/docs/2.4.3/api/R/columnfunctions.html
@@ -0,0 +1,55 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: A set of operations working with SparkDataFrame columns</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for asc {SparkR}"><tr><td>asc {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>A set of operations working with SparkDataFrame columns</h2>
+
+<h3>Description</h3>
+
+<p>A set of operations working with SparkDataFrame columns
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+asc(x)
+
+contains(x, ...)
+
+desc(x)
+
+getField(x, ...)
+
+getItem(x, ...)
+
+isNaN(x)
+
+isNull(x)
+
+isNotNull(x)
+
+like(x, ...)
+
+rlike(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column object.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s).</p>
+</td></tr>
+</table>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/columns.html b/site/docs/2.4.3/api/R/columns.html
new file mode 100644
index 0000000..96c50de
--- /dev/null
+++ b/site/docs/2.4.3/api/R/columns.html
@@ -0,0 +1,145 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Column Names of SparkDataFrame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for colnames {SparkR}"><tr><td>colnames {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Column Names of SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>Return a vector of column names.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+colnames(x, do.NULL = TRUE, prefix = "col")
+
+colnames(x) &lt;- value
+
+columns(x)
+
+## S4 method for signature 'SparkDataFrame'
+columns(x)
+
+## S4 method for signature 'SparkDataFrame'
+names(x)
+
+## S4 replacement method for signature 'SparkDataFrame'
+names(x) &lt;- value
+
+## S4 method for signature 'SparkDataFrame'
+colnames(x)
+
+## S4 replacement method for signature 'SparkDataFrame'
+colnames(x) &lt;- value
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>do.NULL</code></td>
+<td>
+<p>currently not used.</p>
+</td></tr>
+<tr valign="top"><td><code>prefix</code></td>
+<td>
+<p>currently not used.</p>
+</td></tr>
+<tr valign="top"><td><code>value</code></td>
+<td>
+<p>a character vector. Must have the same length as the number
+of columns to be renamed.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>columns since 1.4.0
+</p>
+<p>names since 1.5.0
+</p>
+<p>names&lt;- since 1.5.0
+</p>
+<p>colnames since 1.6.0
+</p>
+<p>colnames&lt;- since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="rename.html">rename</a></code>,
+<code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D columns(df)
+##D colnames(df)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/corr.html b/site/docs/2.4.3/api/R/corr.html
new file mode 100644
index 0000000..554d235
--- /dev/null
+++ b/site/docs/2.4.3/api/R/corr.html
@@ -0,0 +1,110 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: corr</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for corr {SparkR}"><tr><td>corr {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>corr</h2>
+
+<h3>Description</h3>
+
+<p>Computes the Pearson Correlation Coefficient for two Columns.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+corr(x, ...)
+
+## S4 method for signature 'Column'
+corr(x, col2)
+
+## S4 method for signature 'SparkDataFrame'
+corr(x, colName1, colName2,
+  method = "pearson")
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column or a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s). If <code>x</code> is a Column, a Column
+should be provided. If <code>x</code> is a SparkDataFrame, two column names should
+be provided.</p>
+</td></tr>
+<tr valign="top"><td><code>col2</code></td>
+<td>
+<p>a (second) Column.</p>
+</td></tr>
+<tr valign="top"><td><code>colName1</code></td>
+<td>
+<p>the name of the first column</p>
+</td></tr>
+<tr valign="top"><td><code>colName2</code></td>
+<td>
+<p>the name of the second column</p>
+</td></tr>
+<tr valign="top"><td><code>method</code></td>
+<td>
+<p>Optional. A character specifying the method for calculating the correlation.
+only &quot;pearson&quot; is allowed now.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>The Pearson Correlation Coefficient as a Double.
+</p>
+
+
+<h3>Note</h3>
+
+<p>corr since 1.6.0
+</p>
+<p>corr since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other aggregate functions: <code><a href="avg.html">avg</a></code>,
+<code><a href="column_aggregate_functions.html">column_aggregate_functions</a></code>,
+<code><a href="count.html">count</a></code>, <code><a href="cov.html">cov</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="last.html">last</a></code>
+</p>
+<p>Other stat functions: <code><a href="approxQuantile.html">approxQuantile</a></code>,
+<code><a href="cov.html">cov</a></code>, <code><a href="crosstab.html">crosstab</a></code>,
+<code><a href="freqItems.html">freqItems</a></code>, <code><a href="sampleBy.html">sampleBy</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+##D head(select(df, corr(df$mpg, df$hp)))
+## End(Not run)
+
+## Not run: 
+##D corr(df, &quot;mpg&quot;, &quot;hp&quot;)
+##D corr(df, &quot;mpg&quot;, &quot;hp&quot;, method = &quot;pearson&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/count.html b/site/docs/2.4.3/api/R/count.html
new file mode 100644
index 0000000..f2a86b6
--- /dev/null
+++ b/site/docs/2.4.3/api/R/count.html
@@ -0,0 +1,89 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Count</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for count {SparkR}"><tr><td>count {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Count</h2>
+
+<h3>Description</h3>
+
+<p>Count the number of rows for each group when we have <code>GroupedData</code> input.
+The resulting SparkDataFrame will also contain the grouping columns.
+</p>
+<p>This can be used as a column aggregate function with <code>Column</code> as input,
+and returns the number of items in a group.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+count(x)
+
+n(x)
+
+## S4 method for signature 'GroupedData'
+count(x)
+
+## S4 method for signature 'Column'
+count(x)
+
+## S4 method for signature 'Column'
+n(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a GroupedData or Column.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>count since 1.4.0
+</p>
+<p>count since 1.4.0
+</p>
+<p>n since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other aggregate functions: <code><a href="avg.html">avg</a></code>,
+<code><a href="column_aggregate_functions.html">column_aggregate_functions</a></code>,
+<code><a href="corr.html">corr</a></code>, <code><a href="cov.html">cov</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="last.html">last</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D   count(groupBy(df, &quot;name&quot;))
+## End(Not run)
+## Not run: count(df$c)
+## Not run: n(df$c)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/cov.html b/site/docs/2.4.3/api/R/cov.html
new file mode 100644
index 0000000..62decf6
--- /dev/null
+++ b/site/docs/2.4.3/api/R/cov.html
@@ -0,0 +1,137 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: cov</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for cov {SparkR}"><tr><td>cov {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>cov</h2>
+
+<h3>Description</h3>
+
+<p>Compute the covariance between two expressions.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+cov(x, ...)
+
+covar_samp(col1, col2)
+
+covar_pop(col1, col2)
+
+## S4 method for signature 'characterOrColumn'
+cov(x, col2)
+
+## S4 method for signature 'characterOrColumn,characterOrColumn'
+covar_samp(col1, col2)
+
+## S4 method for signature 'characterOrColumn,characterOrColumn'
+covar_pop(col1, col2)
+
+## S4 method for signature 'SparkDataFrame'
+cov(x, colName1, colName2)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column or a SparkDataFrame.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s). If <code>x</code> is a Column, a Column
+should be provided. If <code>x</code> is a SparkDataFrame, two column names should
+be provided.</p>
+</td></tr>
+<tr valign="top"><td><code>col1</code></td>
+<td>
+<p>the first Column.</p>
+</td></tr>
+<tr valign="top"><td><code>col2</code></td>
+<td>
+<p>the second Column.</p>
+</td></tr>
+<tr valign="top"><td><code>colName1</code></td>
+<td>
+<p>the name of the first column</p>
+</td></tr>
+<tr valign="top"><td><code>colName2</code></td>
+<td>
+<p>the name of the second column</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p><code>cov</code>: Compute the sample covariance between two expressions.
+</p>
+<p><code>covar_sample</code>: Alias for <code>cov</code>.
+</p>
+<p><code>covar_pop</code>: Computes the population covariance between two expressions.
+</p>
+<p><code>cov</code>: When applied to SparkDataFrame, this calculates the sample covariance of two
+numerical columns of <em>one</em> SparkDataFrame.
+</p>
+
+
+<h3>Value</h3>
+
+<p>The covariance of the two columns.
+</p>
+
+
+<h3>Note</h3>
+
+<p>cov since 1.6.0
+</p>
+<p>covar_samp since 2.0.0
+</p>
+<p>covar_pop since 2.0.0
+</p>
+<p>cov since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other aggregate functions: <code><a href="avg.html">avg</a></code>,
+<code><a href="column_aggregate_functions.html">column_aggregate_functions</a></code>,
+<code><a href="corr.html">corr</a></code>, <code><a href="count.html">count</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="last.html">last</a></code>
+</p>
+<p>Other stat functions: <code><a href="approxQuantile.html">approxQuantile</a></code>,
+<code><a href="corr.html">corr</a></code>, <code><a href="crosstab.html">crosstab</a></code>,
+<code><a href="freqItems.html">freqItems</a></code>, <code><a href="sampleBy.html">sampleBy</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(cbind(model = rownames(mtcars), mtcars))
+##D head(select(df, cov(df$mpg, df$hp), cov(&quot;mpg&quot;, &quot;hp&quot;),
+##D                 covar_samp(df$mpg, df$hp), covar_samp(&quot;mpg&quot;, &quot;hp&quot;),
+##D                 covar_pop(df$mpg, df$hp), covar_pop(&quot;mpg&quot;, &quot;hp&quot;)))
+## End(Not run)
+
+## Not run: 
+##D cov(df, &quot;mpg&quot;, &quot;hp&quot;)
+##D cov(df, df$mpg, df$hp)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/createDataFrame.html b/site/docs/2.4.3/api/R/createDataFrame.html
new file mode 100644
index 0000000..b39a182
--- /dev/null
+++ b/site/docs/2.4.3/api/R/createDataFrame.html
@@ -0,0 +1,90 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create a SparkDataFrame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for createDataFrame {SparkR}"><tr><td>createDataFrame {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create a SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>Converts R data.frame or list into SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+createDataFrame(data, schema = NULL,
+  samplingRatio = 1, numPartitions = NULL)
+
+## Default S3 method:
+as.DataFrame(data, schema = NULL, samplingRatio = 1,
+  numPartitions = NULL)
+
+as.DataFrame(data, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>data</code></td>
+<td>
+<p>a list or data.frame.</p>
+</td></tr>
+<tr valign="top"><td><code>schema</code></td>
+<td>
+<p>a list of column names or named list (StructType), optional.</p>
+</td></tr>
+<tr valign="top"><td><code>samplingRatio</code></td>
+<td>
+<p>Currently not used.</p>
+</td></tr>
+<tr valign="top"><td><code>numPartitions</code></td>
+<td>
+<p>the number of partitions of the SparkDataFrame. Defaults to 1, this is
+limited by length of the list or number of rows of the data.frame</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s).</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>createDataFrame since 1.4.0
+</p>
+<p>as.DataFrame since 1.6.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df1 &lt;- as.DataFrame(iris)
+##D df2 &lt;- as.DataFrame(list(3,4,5,6))
+##D df3 &lt;- createDataFrame(iris)
+##D df4 &lt;- createDataFrame(cars, numPartitions = 2)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/createExternalTable-deprecated.html b/site/docs/2.4.3/api/R/createExternalTable-deprecated.html
new file mode 100644
index 0000000..5e065b5
--- /dev/null
+++ b/site/docs/2.4.3/api/R/createExternalTable-deprecated.html
@@ -0,0 +1,93 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: (Deprecated) Create an external table</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for createExternalTable {SparkR}"><tr><td>createExternalTable {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Create an external table</h2>
+
+<h3>Description</h3>
+
+<p>Creates an external table based on the dataset in a data source,
+Returns a SparkDataFrame associated with the external table.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+createExternalTable(tableName, path = NULL,
+  source = NULL, schema = NULL, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>a name of the table.</p>
+</td></tr>
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>the path of files to load.</p>
+</td></tr>
+<tr valign="top"><td><code>source</code></td>
+<td>
+<p>the name of external data source.</p>
+</td></tr>
+<tr valign="top"><td><code>schema</code></td>
+<td>
+<p>the schema of the data required for some data sources.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s) passed to the method.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>The data source is specified by the <code>source</code> and a set of options(...).
+If <code>source</code> is not specified, the default data source configured by
+&quot;spark.sql.sources.default&quot; will be used.
+</p>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>createExternalTable since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="createTable.html">createTable</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df &lt;- createExternalTable(&quot;myjson&quot;, path=&quot;path/to/json&quot;, source=&quot;json&quot;, schema)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/createOrReplaceTempView.html b/site/docs/2.4.3/api/R/createOrReplaceTempView.html
new file mode 100644
index 0000000..2ca5712
--- /dev/null
+++ b/site/docs/2.4.3/api/R/createOrReplaceTempView.html
@@ -0,0 +1,112 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Creates a temporary view using the given name.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for createOrReplaceTempView {SparkR}"><tr><td>createOrReplaceTempView {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Creates a temporary view using the given name.</h2>
+
+<h3>Description</h3>
+
+<p>Creates a new temporary view using a SparkDataFrame in the Spark Session. If a
+temporary view with the same name already exists, replaces it.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+createOrReplaceTempView(x, viewName)
+
+## S4 method for signature 'SparkDataFrame,character'
+createOrReplaceTempView(x, viewName)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+<tr valign="top"><td><code>viewName</code></td>
+<td>
+<p>A character vector containing the name of the table</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>createOrReplaceTempView since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>, <code><a href="crossJoin.html">crossJoin</a></code>,
+<code><a href="cube.html">cube</a></code>, <code><a href="dapplyCollect.html">dapplyCollect</a></code>,
+<code><a href="dapply.html">dapply</a></code>, <code><a href="describe.html">describe</a></code>,
+<code><a href="dim.html">dim</a></code>, <code><a href="distinct.html">distinct</a></code>,
+<code><a href="dropDuplicates.html">dropDuplicates</a></code>, <code><a href="nafunctions.html">dropna</a></code>,
+<code><a href="drop.html">drop</a></code>, <code><a href="dtypes.html">dtypes</a></code>,
+<code><a href="exceptAll.html">exceptAll</a></code>, <code><a href="except.html">except</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="filter.html">filter</a></code>,
+<code><a href="first.html">first</a></code>, <code><a href="gapplyCollect.html">gapplyCollect</a></code>,
+<code><a href="gapply.html">gapply</a></code>, <code><a href="getNumPartitions.html">getNumPartitions</a></code>,
+<code><a href="groupBy.html">group_by</a></code>, <code><a href="head.html">head</a></code>,
+<code><a href="hint.html">hint</a></code>, <code><a href="histogram.html">histogram</a></code>,
+<code><a href="insertInto.html">insertInto</a></code>, <code><a href="intersectAll.html">intersectAll</a></code>,
+<code><a href="intersect.html">intersect</a></code>, <code><a href="isLocal.html">isLocal</a></code>,
+<code><a href="isStreaming.html">isStreaming</a></code>, <code><a href="join.html">join</a></code>,
+<code><a href="limit.html">limit</a></code>, <code><a href="localCheckpoint.html">localCheckpoint</a></code>,
+<code><a href="merge.html">merge</a></code>, <code><a href="mutate.html">mutate</a></code>,
+<code><a href="ncol.html">ncol</a></code>, <code><a href="nrow.html">nrow</a></code>,
+<code><a href="persist.html">persist</a></code>, <code><a href="printSchema.html">printSchema</a></code>,
+<code><a href="randomSplit.html">randomSplit</a></code>, <code><a href="rbind.html">rbind</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D createOrReplaceTempView(df, &quot;json_df&quot;)
+##D new_df &lt;- sql(&quot;SELECT * FROM json_df&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.3 <a href="00Index.html">Index</a>]</div>
+</body></html>
diff --git a/site/docs/2.4.3/api/R/createTable.html b/site/docs/2.4.3/api/R/createTable.html
new file mode 100644
index 0000000..cd5c6b0
--- /dev/null
+++ b/site/docs/2.4.3/api/R/createTable.html
@@ -0,0 +1,91 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Creates a table based on the dataset in a data source</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for createTable {SparkR}"><tr><td>createTable {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Creates a table based on the dataset in a data source</h2>
+
+<h3>Description</h3>
+
+<p>Creates a table based on the dataset in a data source. Returns a SparkDataFrame associated with
+the table.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+createTable(tableName, path = NULL, source = NULL, schema = NULL,
+  ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>the qualified or unqualified name that designates a table. If no database
+identifier is provided, it refers to a table in the current database.</p>
+</td></tr>
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>(optional) the path of files to load.</p>
+</td></tr>
+<tr valign="top"><td><code>source</code></td>
+<td>
+<p>(optional) the name of the data source.</p>
+</td></tr>
+<tr valign="top"><td><code>schema</code></td>
+<td>
+<p>(optional) the schema of the data required for some data sources.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional named parameters as options for the data source.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>The data source is specified by the <code>source</code> and a set of options(...).
+If <code>source</code> is not specified, the default data source configured by
+&quot;spark.sql.sources.default&quot; will be used. When a <code>path</code> is specified, an external table is
+created from the data at the given path. Otherwise a managed table is created.
+</p>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame.
+</p>
+
+
+<h3>Note</h3>
+
+<p>createTable since 2.2.0
+</p>
+
+
+<h3>Examples</h3>
... 24940 lines suppressed ...


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org