You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@geode.apache.org by ji...@apache.org on 2016/04/21 19:17:14 UTC

[39/50] [abbrv] incubator-geode git commit: GEODE-1244: Package, directory, project and file rename for geode-spark-connector

http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/Dependencies.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/Dependencies.scala b/geode-spark-connector/project/Dependencies.scala
index 4b6682e..140fdc4 100644
--- a/geode-spark-connector/project/Dependencies.scala
+++ b/geode-spark-connector/project/Dependencies.scala
@@ -22,7 +22,7 @@ object Dependencies {
   object Compile {
     val sparkStreaming = "org.apache.spark" %% "spark-streaming" % "1.3.0" 
     val sparkSql = "org.apache.spark" %% "spark-sql" % "1.3.0"
-    val gemfire = "org.apache.geode" % "geode-core" % "1.0.0-incubating.M3-SNAPSHOT" excludeAll(ExclusionRule(organization = "org.jboss.netty") )
+    val geode = "org.apache.geode" % "geode-core" % "1.0.0-incubating.M3-SNAPSHOT" excludeAll(ExclusionRule(organization = "org.jboss.netty") )
   }
 
   object Test {
@@ -37,9 +37,9 @@ object Dependencies {
 
   val unitTests = Seq(scalaTest, mockito, junit, novoCode)
 
-  val connector = unitTests ++ Seq(sparkStreaming, sparkSql, gemfire)
+  val connector = unitTests ++ Seq(sparkStreaming, sparkSql, geode)
 
-  val functions = Seq(gemfire, junit)
+  val functions = Seq(geode, junit)
  
-  val demos = Seq(sparkStreaming, sparkSql, gemfire)
+  val demos = Seq(sparkStreaming, sparkSql, geode)
 }

http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/GemFireSparkBuild.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/GemFireSparkBuild.scala b/geode-spark-connector/project/GemFireSparkBuild.scala
deleted file mode 100644
index 89d8e0b..0000000
--- a/geode-spark-connector/project/GemFireSparkBuild.scala
+++ /dev/null
@@ -1,76 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *      http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-import sbt._
-import sbt.Keys._
-import scoverage.ScoverageSbtPlugin._
-import scoverage.ScoverageSbtPlugin
-
-object GemFireSparkConnectorBuild extends Build {
-  import Settings._
-  import Dependencies._ 
-
-  lazy val root = Project(
-    id = "root", 
-    base =file("."), 
-    aggregate = Seq(gemfireFunctions, gemfireSparkConnector,demos),
-    settings = commonSettings ++ Seq( 
-     name := "GemFire Connector for Apache Spark",
-     publishArtifact :=  false,
-     publishLocal := { },
-     publish := { }
-    )
-  )
- 
-  lazy val gemfireFunctions = Project(
-    id = "gemfire-functions",
-    base = file("gemfire-functions"),
-    settings = commonSettings ++ Seq(libraryDependencies ++= Dependencies.functions,
-      resolvers ++= gfcResolvers,
-      description := "Required GemFire Functions to be deployed onto the GemFire Cluster before using the GemFire Spark Connector"
-    )
-  ).configs(IntegrationTest)
-  
-  lazy val gemfireSparkConnector = Project(
-    id = "gemfire-spark-connector",
-    base = file("gemfire-spark-connector"),
-    settings = gfcSettings ++ Seq(libraryDependencies ++= Dependencies.connector,
-      resolvers ++= gfcResolvers,
-      description := "A library that exposes GemFire regions as Spark RDDs, writes Spark RDDs to GemFire regions, and executes OQL queries from Spark Applications to GemFire"
-    )
-  ).dependsOn(gemfireFunctions).configs(IntegrationTest)
-
- 
-  /******** Demo Project Definitions ********/ 
-  lazy val demoPath = file("gemfire-spark-demos")
-
-  lazy val demos = Project ( 
-    id = "gemfire-spark-demos",
-    base = demoPath,
-    settings = demoSettings,
-    aggregate = Seq(basicDemos)
-  )
- 
-  lazy val basicDemos = Project (
-    id = "basic-demos",
-    base = demoPath / "basic-demos",
-    settings = demoSettings ++ Seq(libraryDependencies ++= Dependencies.demos,
-      resolvers ++= gfcResolvers,
-      description := "Sample applications that demonstrates functionality of the GemFire Spark Connector"
-    )
-  ).dependsOn(gemfireSparkConnector)
-}
-

http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/GeodeSparkBuild.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/GeodeSparkBuild.scala b/geode-spark-connector/project/GeodeSparkBuild.scala
new file mode 100644
index 0000000..07cae51
--- /dev/null
+++ b/geode-spark-connector/project/GeodeSparkBuild.scala
@@ -0,0 +1,76 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *      http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+import sbt._
+import sbt.Keys._
+import scoverage.ScoverageSbtPlugin._
+import scoverage.ScoverageSbtPlugin
+
+object GeodeSparkConnectorBuild extends Build {
+  import Settings._
+  import Dependencies._ 
+
+  lazy val root = Project(
+    id = "root", 
+    base =file("."), 
+    aggregate = Seq(geodeFunctions, geodeSparkConnector,demos),
+    settings = commonSettings ++ Seq( 
+     name := "Geode Connector for Apache Spark",
+     publishArtifact :=  false,
+     publishLocal := { },
+     publish := { }
+    )
+  )
+ 
+  lazy val geodeFunctions = Project(
+    id = "geode-functions",
+    base = file("geode-functions"),
+    settings = commonSettings ++ Seq(libraryDependencies ++= Dependencies.functions,
+      resolvers ++= gfcResolvers,
+      description := "Required Geode Functions to be deployed onto the Geode Cluster before using the Geode Spark Connector"
+    )
+  ).configs(IntegrationTest)
+  
+  lazy val geodeSparkConnector = Project(
+    id = "geode-spark-connector",
+    base = file("geode-spark-connector"),
+    settings = gfcSettings ++ Seq(libraryDependencies ++= Dependencies.connector,
+      resolvers ++= gfcResolvers,
+      description := "A library that exposes Geode regions as Spark RDDs, writes Spark RDDs to Geode regions, and executes OQL queries from Spark Applications to Geode"
+    )
+  ).dependsOn(geodeFunctions).configs(IntegrationTest)
+
+ 
+  /******** Demo Project Definitions ********/ 
+  lazy val demoPath = file("geode-spark-demos")
+
+  lazy val demos = Project ( 
+    id = "geode-spark-demos",
+    base = demoPath,
+    settings = demoSettings,
+    aggregate = Seq(basicDemos)
+  )
+ 
+  lazy val basicDemos = Project (
+    id = "basic-demos",
+    base = demoPath / "basic-demos",
+    settings = demoSettings ++ Seq(libraryDependencies ++= Dependencies.demos,
+      resolvers ++= gfcResolvers,
+      description := "Sample applications that demonstrates functionality of the Geode Spark Connector"
+    )
+  ).dependsOn(geodeSparkConnector)
+}
+

http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/Settings.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/Settings.scala b/geode-spark-connector/project/Settings.scala
index 796541c..9aefa9b 100644
--- a/geode-spark-connector/project/Settings.scala
+++ b/geode-spark-connector/project/Settings.scala
@@ -24,7 +24,7 @@ object Settings extends Build {
     organization := "io.pivotal",
     version := "0.5.0",
     scalaVersion := "2.10.4",
-    organization := "io.pivotal.gemfire.spark",
+    organization := "io.pivotal.geode.spark",
     organizationHomepage := Some(url("http://www.pivotal.io/"))
   ) 
 
@@ -43,7 +43,7 @@ object Settings extends Build {
   val gfcITSettings = inConfig(IntegrationTest)(Defaults.itSettings) ++
     Seq(parallelExecution in IntegrationTest := false, fork in IntegrationTest := true)
 
-  val gfcCompileSettings = inConfig(Compile)(Defaults.compileSettings) ++ Seq(unmanagedSourceDirectories in Compile += baseDirectory.value /"../gemfire-functions/src")
+  val gfcCompileSettings = inConfig(Compile)(Defaults.compileSettings) ++ Seq(unmanagedSourceDirectories in Compile += baseDirectory.value /"../geode-functions/src")
 
   val gfcSettings = commonSettings ++ gfcITSettings ++ gfcCompileSettings