You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by ke...@apache.org on 2016/10/31 17:39:46 UTC
[2/3] incubator-beam git commit: Fixes a compile error in SourceRDD
Fixes a compile error in SourceRDD
Project: http://git-wip-us.apache.org/repos/asf/incubator-beam/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-beam/commit/63cbb974
Tree: http://git-wip-us.apache.org/repos/asf/incubator-beam/tree/63cbb974
Diff: http://git-wip-us.apache.org/repos/asf/incubator-beam/diff/63cbb974
Branch: refs/heads/master
Commit: 63cbb9742d836dff250ba6d8d447dc54d5b76a54
Parents: 54a7374
Author: Eugene Kirpichov <ki...@google.com>
Authored: Fri Oct 28 15:21:00 2016 -0700
Committer: Kenneth Knowles <kl...@google.com>
Committed: Mon Oct 31 10:25:48 2016 -0700
----------------------------------------------------------------------
.../main/java/org/apache/beam/runners/spark/io/SourceRDD.java | 4 +++-
1 file changed, 3 insertions(+), 1 deletion(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/incubator-beam/blob/63cbb974/runners/spark/src/main/java/org/apache/beam/runners/spark/io/SourceRDD.java
----------------------------------------------------------------------
diff --git a/runners/spark/src/main/java/org/apache/beam/runners/spark/io/SourceRDD.java b/runners/spark/src/main/java/org/apache/beam/runners/spark/io/SourceRDD.java
index 3995c89..cf37b3a 100644
--- a/runners/spark/src/main/java/org/apache/beam/runners/spark/io/SourceRDD.java
+++ b/runners/spark/src/main/java/org/apache/beam/runners/spark/io/SourceRDD.java
@@ -216,8 +216,10 @@ public class SourceRDD {
private final MicrobatchSource<T, CheckpointMarkT> microbatchSource;
private final SparkRuntimeContext runtimeContext;
+ // to satisfy Scala API.
private static final scala.collection.immutable.List<Dependency<?>> NIL =
- scala.collection.immutable.List.empty();
+ scala.collection.JavaConversions
+ .asScalaBuffer(Collections.<Dependency<?>>emptyList()).toList();
public Unbounded(SparkContext sc,
SparkRuntimeContext runtimeContext,