You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@sedona.apache.org by ji...@apache.org on 2021/03/29 00:14:47 UTC
[incubator-sedona] branch master updated: Add Scala Adapter Test
for HDF data
This is an automated email from the ASF dual-hosted git repository.
jiayu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-sedona.git
The following commit(s) were added to refs/heads/master by this push:
new 6169c5c Add Scala Adapter Test for HDF data
6169c5c is described below
commit 6169c5c9396daad1b568f109753250f57c206d7f
Author: Jia Yu <ji...@apache.org>
AuthorDate: Sun Mar 28 17:14:38 2021 -0700
Add Scala Adapter Test for HDF data
---
.../org/apache/sedona/sql/adapterTestScala.scala | 20 +++++++++++++++++++-
1 file changed, 19 insertions(+), 1 deletion(-)
diff --git a/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala b/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
index b4cc427..2248b78 100644
--- a/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
+++ b/sql/src/test/scala/org/apache/sedona/sql/adapterTestScala.scala
@@ -20,11 +20,13 @@
package org.apache.sedona.sql
import org.apache.sedona.core.enums.{FileDataSplitter, GridType, IndexType}
+import org.apache.sedona.core.formatMapper.EarthdataHDFPointMapper
import org.apache.sedona.core.formatMapper.shapefileParser.ShapefileReader
import org.apache.sedona.core.spatialOperator.JoinQuery
-import org.apache.sedona.core.spatialRDD.{CircleRDD, PolygonRDD}
+import org.apache.sedona.core.spatialRDD.{CircleRDD, PointRDD, PolygonRDD}
import org.apache.sedona.sql.utils.Adapter
import org.apache.spark.sql.sedona_sql.UDT.GeometryUDT
+import org.apache.spark.storage.StorageLevel
import org.scalatest.GivenWhenThen
import org.scalatest.matchers.should.Matchers
@@ -178,7 +180,23 @@ class adapterTestScala extends TestBaseScala with Matchers with GivenWhenThen{
val df = Adapter.toDf(spatialRDD, sparkSession)
assert(df.columns.length == 4)
assert(df.count() == 1)
+ }
+ it("load HDF data from RDD to a DataFrame") {
+ val InputLocation = "file://" + resourceFolder + "modis/modis.csv"
+ val numPartitions = 5
+ val HDFincrement = 5
+ val HDFoffset = 2
+ val HDFrootGroupName = "MOD_Swath_LST"
+ val HDFDataVariableName = "LST"
+ val urlPrefix = resourceFolder + "modis/"
+ val HDFDataVariableList:Array[String] = Array("LST", "QC", "Error_LST", "Emis_31", "Emis_32")
+ val earthdataHDFPoint = new EarthdataHDFPointMapper(HDFincrement, HDFoffset, HDFrootGroupName, HDFDataVariableList, HDFDataVariableName, urlPrefix)
+ val spatialRDD = new PointRDD(sparkSession.sparkContext, InputLocation, numPartitions, earthdataHDFPoint, StorageLevel.MEMORY_ONLY)
+ import scala.collection.JavaConverters._
+ spatialRDD.fieldNames = HDFDataVariableList.dropRight(4).toList.asJava
+ val spatialDf = Adapter.toDf(spatialRDD, sparkSession)
+ assert(spatialDf.schema.fields(1).name == "LST")
}
}
}