You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2014/09/17 01:52:12 UTC
svn commit: r1625435 - in
/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark:
GenSparkProcContext.java GenSparkUtils.java
Author: xuefu
Date: Tue Sep 16 23:52:12 2014
New Revision: 1625435
URL: http://svn.apache.org/r1625435
Log:
HIVE-8055: Code cleanup after HIVE-8054 [Spark Branch] (Na via Xuefu)
Modified:
hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkProcContext.java
hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkProcContext.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkProcContext.java?rev=1625435&r1=1625434&r2=1625435&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkProcContext.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkProcContext.java Tue Sep 16 23:52:12 2014
@@ -26,7 +26,6 @@ import java.util.List;
import java.util.Map;
import java.util.Set;
-import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.exec.DependencyCollectionTask;
import org.apache.hadoop.hive.ql.exec.FileSinkOperator;
@@ -45,7 +44,6 @@ import org.apache.hadoop.hive.ql.plan.Ba
import org.apache.hadoop.hive.ql.plan.DependencyCollectionWork;
import org.apache.hadoop.hive.ql.plan.MoveWork;
import org.apache.hadoop.hive.ql.plan.OperatorDesc;
-import org.apache.hadoop.hive.ql.plan.FileSinkDesc;
import org.apache.hadoop.hive.ql.plan.SparkEdgeProperty;
import org.apache.hadoop.hive.ql.plan.SparkWork;
@@ -125,8 +123,6 @@ public class GenSparkProcContext impleme
public final Set<BaseWork> workWithUnionOperators;
public final Set<ReduceSinkOperator> clonedReduceSinks;
- // we link filesink that will write to the same final location
- public final Map<Path, List<FileSinkDesc>> linkedFileSinks;
public final Set<FileSinkOperator> fileSinkSet;
// remember which reducesinks we've already connected
@@ -159,7 +155,6 @@ public class GenSparkProcContext impleme
this.currentUnionOperators = new LinkedList<UnionOperator>();
this.workWithUnionOperators = new LinkedHashSet<BaseWork>();
this.clonedReduceSinks = new LinkedHashSet<ReduceSinkOperator>();
- this.linkedFileSinks = new LinkedHashMap<Path, List<FileSinkDesc>>();
this.fileSinkSet = new LinkedHashSet<FileSinkOperator>();
this.connectedReduceSinks = new LinkedHashSet<ReduceSinkOperator>();
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java?rev=1625435&r1=1625434&r2=1625435&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java Tue Sep 16 23:52:12 2014
@@ -213,22 +213,6 @@ public class GenSparkUtils {
Operator<?> current = operators.pop();
seen.add(current);
- if (current instanceof FileSinkOperator) {
- FileSinkOperator fileSink = (FileSinkOperator)current;
-
- FileSinkDesc desc = fileSink.getConf();
- Path path = desc.getDirName();
- List<FileSinkDesc> linked;
-
- if (!context.linkedFileSinks.containsKey(path)) {
- linked = new ArrayList<FileSinkDesc>();
- context.linkedFileSinks.put(path, linked);
- }
- linked = context.linkedFileSinks.get(path);
- linked.add(desc);
- desc.setLinkedFileSinkDesc(linked);
- }
-
if (current instanceof UnionOperator) {
Operator<?> parent = null;
int count = 0;