You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@sedona.apache.org by ji...@apache.org on 2023/02/10 08:11:07 UTC
[sedona] branch master updated: [DOCS] Fix spelling (#759)
This is an automated email from the ASF dual-hosted git repository.
jiayu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/sedona.git
The following commit(s) were added to refs/heads/master by this push:
new de2d2d84 [DOCS] Fix spelling (#759)
de2d2d84 is described below
commit de2d2d84d9c1339fc91d6e94edc48b2f03116cdf
Author: John Bampton <jb...@users.noreply.github.com>
AuthorDate: Fri Feb 10 18:11:00 2023 +1000
[DOCS] Fix spelling (#759)
---
R/R/data_interface.R | 22 +++++++++++-----------
mkdocs.yml | 2 +-
python/sedona/core/SpatialRDD/spatial_rdd.py | 2 +-
.../datasources/parquet/GeoDateTimeUtils.scala | 2 +-
.../sedona_sql/expressions/raster/Functions.scala | 2 +-
.../org/apache/sedona/sql/adapterTestScala.scala | 2 +-
6 files changed, 16 insertions(+), 16 deletions(-)
diff --git a/R/R/data_interface.R b/R/R/data_interface.R
index d2ce6d47..436da7bb 100644
--- a/R/R/data_interface.R
+++ b/R/R/data_interface.R
@@ -71,7 +71,7 @@ NULL
#' )
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_dsv_to_typed_rdd <- function(sc,
@@ -169,7 +169,7 @@ sedona_read_dsv_to_typed_rdd <- function(sc,
#' )
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_shapefile_to_typed_rdd <- function(sc,
@@ -210,7 +210,7 @@ sedona_read_shapefile_to_typed_rdd <- function(sc,
#' )
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_geojson_to_typed_rdd <- function(sc,
@@ -256,7 +256,7 @@ sedona_read_geojson_to_typed_rdd <- function(sc,
#' rdd <- sedona_read_geojson(sc, location = input_location)
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_geojson <- function(sc,
@@ -313,7 +313,7 @@ sedona_read_geojson <- function(sc,
#' )
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_wkb <- function(sc,
@@ -373,7 +373,7 @@ sedona_read_wkb <- function(sc,
#' )
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_wkt <- function(sc,
@@ -422,7 +422,7 @@ sedona_read_wkt <- function(sc,
#' rdd <- sedona_read_shapefile(sc, location = input_location)
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_read_shapefile <- function(sc,
@@ -473,7 +473,7 @@ NULL
#' sedona_write_wkb(rdd, "/tmp/wkb_output.tsv")
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_write_wkb <- function(x, output_location) {
@@ -504,7 +504,7 @@ sedona_write_wkb <- function(x, output_location) {
#' sedona_write_wkt(rdd, "/tmp/wkt_output.tsv")
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_write_wkt <- function(x, output_location) {
@@ -534,7 +534,7 @@ sedona_write_wkt <- function(x, output_location) {
#' sedona_write_geojson(rdd, "/tmp/example.json")
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_write_geojson <- function(x, output_location) {
@@ -573,7 +573,7 @@ sedona_write_geojson <- function(x, output_location) {
#' )
#' }
#'
-#' @family Sedona data inferface functions
+#' @family Sedona data interface functions
#'
#' @export
sedona_save_spatial_rdd <- function(x,
diff --git a/mkdocs.yml b/mkdocs.yml
index c5a2bae9..2834d5dc 100644
--- a/mkdocs.yml
+++ b/mkdocs.yml
@@ -18,7 +18,7 @@ nav:
- Install Sedona R: setup/install-r.md
- Install Sedona-Zeppelin: setup/zeppelin.md
- Install on Databricks: setup/databricks.md
- - Set up Spark cluser: setup/cluster.md
+ - Set up Spark cluster: setup/cluster.md
- Install with Apache Flink:
- Install Sedona Scala/Java: setup/flink/install-scala.md
- Release notes: setup/release-notes.md
diff --git a/python/sedona/core/SpatialRDD/spatial_rdd.py b/python/sedona/core/SpatialRDD/spatial_rdd.py
index 3a51c2c1..429d1000 100644
--- a/python/sedona/core/SpatialRDD/spatial_rdd.py
+++ b/python/sedona/core/SpatialRDD/spatial_rdd.py
@@ -114,7 +114,7 @@ class SpatialRDD:
def CRSTransform(self, sourceEpsgCRSCode: crs, targetEpsgCRSCode: crs) -> bool:
"""
Function transforms coordinates from one crs to another one
- :param sourceEpsgCRSCode: crs, Cooridnate Reference System to transform from
+ :param sourceEpsgCRSCode: crs, Coordinate Reference System to transform from
:param targetEpsgCRSCode: crs, Coordinate Reference System to transform to
:return: bool, True if transforming was correct
"""
diff --git a/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala b/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala
index 7a661542..9cfd71ce 100644
--- a/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala
+++ b/sql/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/GeoDateTimeUtils.scala
@@ -23,7 +23,7 @@ object GeoDateTimeUtils {
* precision, so this conversion is lossy.
*/
def microsToMillis(micros: Long): Long = {
- // When the timestamp is negative i.e before 1970, we need to adjust the millseconds portion.
+ // When the timestamp is negative i.e before 1970, we need to adjust the milliseconds portion.
// Example - 1965-01-01 10:11:12.123456 is represented as (-157700927876544) in micro precision.
// In millis precision the above needs to be represented as (-157700927877).
Math.floorDiv(micros, MICROS_PER_MILLIS)
diff --git a/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala b/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala
index 84732416..0c9a80f7 100644
--- a/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala
+++ b/sql/src/main/scala/org/apache/spark/sql/sedona_sql/expressions/raster/Functions.scala
@@ -315,7 +315,7 @@ case class RS_LessThanEqual(inputExpressions: Seq[Expression])
}
}
-// Count number of occurences of a particular value in a band
+// Count number of occurrences of a particular value in a band
case class RS_Count(inputExpressions: Seq[Expression])
extends Expression with CodegenFallback with UserDataGeneratator {
assert(inputExpressions.length == 2)
diff --git a/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala b/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
index 43dab57d..9f236812 100644
--- a/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
+++ b/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
@@ -413,7 +413,7 @@ class adapterTestScala extends TestBaseScala with GivenWhenThen{
// Check results
// Force an action so that spark has to serialize the data -- this will surface
- // a serialization error if the schema or coersion is incorrect, e.g.
+ // a serialization error if the schema or coercion is incorrect, e.g.
// "Error while encoding: java.lang.RuntimeException: <desired data type> is not a
// valid external type for schema of <current data type>"
println(joinResultDf.show(1))