You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@sedona.apache.org by ji...@apache.org on 2023/02/10 08:11:07 UTC

[sedona] branch master updated: [DOCS] Fix spelling (#759)

This is an automated email from the ASF dual-hosted git repository.

jiayu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/sedona.git


The following commit(s) were added to refs/heads/master by this push:
     new de2d2d84 [DOCS] Fix spelling (#759)
de2d2d84 is described below

commit de2d2d84d9c1339fc91d6e94edc48b2f03116cdf
Author: John Bampton <jb...@users.noreply.github.com>
AuthorDate: Fri Feb 10 18:11:00 2023 +1000

    [DOCS] Fix spelling (#759)
---
 R/R/data_interface.R                               | 22 +++++++++++-----------
 mkdocs.yml                                         |  2 +-
 python/sedona/core/SpatialRDD/spatial_rdd.py       |  2 +-
 .../datasources/parquet/GeoDateTimeUtils.scala     |  2 +-
 .../sedona_sql/expressions/raster/Functions.scala  |  2 +-
 .../org/apache/sedona/sql/adapterTestScala.scala   |  2 +-
 6 files changed, 16 insertions(+), 16 deletions(-)

diff --git a/R/R/data_interface.R b/R/R/data_interface.R
index d2ce6d47..436da7bb 100644
--- a/R/R/data_interface.R
+++ b/R/R/data_interface.R
@@ -71,7 +71,7 @@ NULL
 #'   )
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_dsv_to_typed_rdd <- function(sc,
@@ -169,7 +169,7 @@ sedona_read_dsv_to_typed_rdd <- function(sc,
 #'   )
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_shapefile_to_typed_rdd <- function(sc,
@@ -210,7 +210,7 @@ sedona_read_shapefile_to_typed_rdd <- function(sc,
 #'   )
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_geojson_to_typed_rdd <- function(sc,
@@ -256,7 +256,7 @@ sedona_read_geojson_to_typed_rdd <- function(sc,
 #'   rdd <- sedona_read_geojson(sc, location = input_location)
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_geojson <- function(sc,
@@ -313,7 +313,7 @@ sedona_read_geojson <- function(sc,
 #'   )
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_wkb <- function(sc,
@@ -373,7 +373,7 @@ sedona_read_wkb <- function(sc,
 #'   )
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_wkt <- function(sc,
@@ -422,7 +422,7 @@ sedona_read_wkt <- function(sc,
 #'   rdd <- sedona_read_shapefile(sc, location = input_location)
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_read_shapefile <- function(sc,
@@ -473,7 +473,7 @@ NULL
 #'   sedona_write_wkb(rdd, "/tmp/wkb_output.tsv")
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_write_wkb <- function(x, output_location) {
@@ -504,7 +504,7 @@ sedona_write_wkb <- function(x, output_location) {
 #'   sedona_write_wkt(rdd, "/tmp/wkt_output.tsv")
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_write_wkt <- function(x, output_location) {
@@ -534,7 +534,7 @@ sedona_write_wkt <- function(x, output_location) {
 #'   sedona_write_geojson(rdd, "/tmp/example.json")
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_write_geojson <- function(x, output_location) {
@@ -573,7 +573,7 @@ sedona_write_geojson <- function(x, output_location) {
 #'   )
 #' }
 #'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
 #'
 #' @export
 sedona_save_spatial_rdd <- function(x,
diff --git a/mkdocs.yml b/mkdocs.yml
index c5a2bae9..2834d5dc 100644
--- a/mkdocs.yml
+++ b/mkdocs.yml
@@ -18,7 +18,7 @@ nav:
         - Install Sedona R: setup/install-r.md
         - Install Sedona-Zeppelin: setup/zeppelin.md
         - Install on Databricks: setup/databricks.md
-        - Set up Spark cluser: setup/cluster.md
+        - Set up Spark cluster: setup/cluster.md
       - Install with Apache Flink:
         - Install Sedona Scala/Java: setup/flink/install-scala.md
       - Release notes: setup/release-notes.md      
diff --git a/python/sedona/core/SpatialRDD/spatial_rdd.py b/python/sedona/core/SpatialRDD/spatial_rdd.py
index 3a51c2c1..429d1000 100644
--- a/python/sedona/core/SpatialRDD/spatial_rdd.py
+++ b/python/sedona/core/SpatialRDD/spatial_rdd.py
@@ -114,7 +114,7 @@ class SpatialRDD:
     def CRSTransform(self, sourceEpsgCRSCode: crs, targetEpsgCRSCode: crs) -> bool:
         """
         Function transforms coordinates from one crs to another one
-        :param sourceEpsgCRSCode: crs,  Cooridnate Reference System to transform from
+        :param sourceEpsgCRSCode: crs,  Coordinate Reference System to transform from
         :param targetEpsgCRSCode: crs, Coordinate Reference System to transform to
         :return: bool, True if transforming was correct
         """
diff --git a/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala b/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala
index 7a661542..9cfd71ce 100644
--- a/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala
+++ b/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala
@@ -23,7 +23,7 @@ object GeoDateTimeUtils {
    * precision, so this conversion is lossy.
    */
   def microsToMillis(micros: Long): Long = {
-    // When the timestamp is negative i.e before 1970, we need to adjust the millseconds portion.
+    // When the timestamp is negative i.e before 1970, we need to adjust the milliseconds portion.
     // Example - 1965-01-01 10:11:12.123456 is represented as (-157700927876544) in micro precision.
     // In millis precision the above needs to be represented as (-157700927877).
     Math.floorDiv(micros, MICROS_PER_MILLIS)
diff --git a/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala b/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala
index 84732416..0c9a80f7 100644
--- a/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala
+++ b/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala
@@ -315,7 +315,7 @@ case class RS_LessThanEqual(inputExpressions: Seq[Expression])
   }
 }
 
-// Count number of occurences of a particular value in a band
+// Count number of occurrences of a particular value in a band
 case class RS_Count(inputExpressions: Seq[Expression])
   extends Expression with CodegenFallback with UserDataGeneratator {
   assert(inputExpressions.length == 2)
diff --git a/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala b/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
index 43dab57d..9f236812 100644
--- a/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
+++ b/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
@@ -413,7 +413,7 @@ class adapterTestScala extends TestBaseScala with GivenWhenThen{
 
       // Check results
       // Force an action so that spark has to serialize the data -- this will surface
-      // a serialization error if the schema or coersion is incorrect, e.g.
+      // a serialization error if the schema or coercion is incorrect, e.g.
       // "Error while encoding: java.lang.RuntimeException: <desired data type> is not a
       // valid external type for schema of <current data type>"
       println(joinResultDf.show(1))