You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@geode.apache.org by ji...@apache.org on 2016/04/21 19:17:14 UTC
[39/50] [abbrv] incubator-geode git commit: GEODE-1244: Package,
directory, project and file rename for geode-spark-connector
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/Dependencies.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/Dependencies.scala b/geode-spark-connector/project/Dependencies.scala
index 4b6682e..140fdc4 100644
--- a/geode-spark-connector/project/Dependencies.scala
+++ b/geode-spark-connector/project/Dependencies.scala
@@ -22,7 +22,7 @@ object Dependencies {
object Compile {
val sparkStreaming = "org.apache.spark" %% "spark-streaming" % "1.3.0"
val sparkSql = "org.apache.spark" %% "spark-sql" % "1.3.0"
- val gemfire = "org.apache.geode" % "geode-core" % "1.0.0-incubating.M3-SNAPSHOT" excludeAll(ExclusionRule(organization = "org.jboss.netty") )
+ val geode = "org.apache.geode" % "geode-core" % "1.0.0-incubating.M3-SNAPSHOT" excludeAll(ExclusionRule(organization = "org.jboss.netty") )
}
object Test {
@@ -37,9 +37,9 @@ object Dependencies {
val unitTests = Seq(scalaTest, mockito, junit, novoCode)
- val connector = unitTests ++ Seq(sparkStreaming, sparkSql, gemfire)
+ val connector = unitTests ++ Seq(sparkStreaming, sparkSql, geode)
- val functions = Seq(gemfire, junit)
+ val functions = Seq(geode, junit)
- val demos = Seq(sparkStreaming, sparkSql, gemfire)
+ val demos = Seq(sparkStreaming, sparkSql, geode)
}
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/GemFireSparkBuild.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/GemFireSparkBuild.scala b/geode-spark-connector/project/GemFireSparkBuild.scala
deleted file mode 100644
index 89d8e0b..0000000
--- a/geode-spark-connector/project/GemFireSparkBuild.scala
+++ /dev/null
@@ -1,76 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-import sbt._
-import sbt.Keys._
-import scoverage.ScoverageSbtPlugin._
-import scoverage.ScoverageSbtPlugin
-
-object GemFireSparkConnectorBuild extends Build {
- import Settings._
- import Dependencies._
-
- lazy val root = Project(
- id = "root",
- base =file("."),
- aggregate = Seq(gemfireFunctions, gemfireSparkConnector,demos),
- settings = commonSettings ++ Seq(
- name := "GemFire Connector for Apache Spark",
- publishArtifact := false,
- publishLocal := { },
- publish := { }
- )
- )
-
- lazy val gemfireFunctions = Project(
- id = "gemfire-functions",
- base = file("gemfire-functions"),
- settings = commonSettings ++ Seq(libraryDependencies ++= Dependencies.functions,
- resolvers ++= gfcResolvers,
- description := "Required GemFire Functions to be deployed onto the GemFire Cluster before using the GemFire Spark Connector"
- )
- ).configs(IntegrationTest)
-
- lazy val gemfireSparkConnector = Project(
- id = "gemfire-spark-connector",
- base = file("gemfire-spark-connector"),
- settings = gfcSettings ++ Seq(libraryDependencies ++= Dependencies.connector,
- resolvers ++= gfcResolvers,
- description := "A library that exposes GemFire regions as Spark RDDs, writes Spark RDDs to GemFire regions, and executes OQL queries from Spark Applications to GemFire"
- )
- ).dependsOn(gemfireFunctions).configs(IntegrationTest)
-
-
- /******** Demo Project Definitions ********/
- lazy val demoPath = file("gemfire-spark-demos")
-
- lazy val demos = Project (
- id = "gemfire-spark-demos",
- base = demoPath,
- settings = demoSettings,
- aggregate = Seq(basicDemos)
- )
-
- lazy val basicDemos = Project (
- id = "basic-demos",
- base = demoPath / "basic-demos",
- settings = demoSettings ++ Seq(libraryDependencies ++= Dependencies.demos,
- resolvers ++= gfcResolvers,
- description := "Sample applications that demonstrates functionality of the GemFire Spark Connector"
- )
- ).dependsOn(gemfireSparkConnector)
-}
-
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/GeodeSparkBuild.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/GeodeSparkBuild.scala b/geode-spark-connector/project/GeodeSparkBuild.scala
new file mode 100644
index 0000000..07cae51
--- /dev/null
+++ b/geode-spark-connector/project/GeodeSparkBuild.scala
@@ -0,0 +1,76 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+import sbt._
+import sbt.Keys._
+import scoverage.ScoverageSbtPlugin._
+import scoverage.ScoverageSbtPlugin
+
+object GeodeSparkConnectorBuild extends Build {
+ import Settings._
+ import Dependencies._
+
+ lazy val root = Project(
+ id = "root",
+ base =file("."),
+ aggregate = Seq(geodeFunctions, geodeSparkConnector,demos),
+ settings = commonSettings ++ Seq(
+ name := "Geode Connector for Apache Spark",
+ publishArtifact := false,
+ publishLocal := { },
+ publish := { }
+ )
+ )
+
+ lazy val geodeFunctions = Project(
+ id = "geode-functions",
+ base = file("geode-functions"),
+ settings = commonSettings ++ Seq(libraryDependencies ++= Dependencies.functions,
+ resolvers ++= gfcResolvers,
+ description := "Required Geode Functions to be deployed onto the Geode Cluster before using the Geode Spark Connector"
+ )
+ ).configs(IntegrationTest)
+
+ lazy val geodeSparkConnector = Project(
+ id = "geode-spark-connector",
+ base = file("geode-spark-connector"),
+ settings = gfcSettings ++ Seq(libraryDependencies ++= Dependencies.connector,
+ resolvers ++= gfcResolvers,
+ description := "A library that exposes Geode regions as Spark RDDs, writes Spark RDDs to Geode regions, and executes OQL queries from Spark Applications to Geode"
+ )
+ ).dependsOn(geodeFunctions).configs(IntegrationTest)
+
+
+ /******** Demo Project Definitions ********/
+ lazy val demoPath = file("geode-spark-demos")
+
+ lazy val demos = Project (
+ id = "geode-spark-demos",
+ base = demoPath,
+ settings = demoSettings,
+ aggregate = Seq(basicDemos)
+ )
+
+ lazy val basicDemos = Project (
+ id = "basic-demos",
+ base = demoPath / "basic-demos",
+ settings = demoSettings ++ Seq(libraryDependencies ++= Dependencies.demos,
+ resolvers ++= gfcResolvers,
+ description := "Sample applications that demonstrates functionality of the Geode Spark Connector"
+ )
+ ).dependsOn(geodeSparkConnector)
+}
+
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/ff914bd9/geode-spark-connector/project/Settings.scala
----------------------------------------------------------------------
diff --git a/geode-spark-connector/project/Settings.scala b/geode-spark-connector/project/Settings.scala
index 796541c..9aefa9b 100644
--- a/geode-spark-connector/project/Settings.scala
+++ b/geode-spark-connector/project/Settings.scala
@@ -24,7 +24,7 @@ object Settings extends Build {
organization := "io.pivotal",
version := "0.5.0",
scalaVersion := "2.10.4",
- organization := "io.pivotal.gemfire.spark",
+ organization := "io.pivotal.geode.spark",
organizationHomepage := Some(url("http://www.pivotal.io/"))
)
@@ -43,7 +43,7 @@ object Settings extends Build {
val gfcITSettings = inConfig(IntegrationTest)(Defaults.itSettings) ++
Seq(parallelExecution in IntegrationTest := false, fork in IntegrationTest := true)
- val gfcCompileSettings = inConfig(Compile)(Defaults.compileSettings) ++ Seq(unmanagedSourceDirectories in Compile += baseDirectory.value /"../gemfire-functions/src")
+ val gfcCompileSettings = inConfig(Compile)(Defaults.compileSettings) ++ Seq(unmanagedSourceDirectories in Compile += baseDirectory.value /"../geode-functions/src")
val gfcSettings = commonSettings ++ gfcITSettings ++ gfcCompileSettings