You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2014/10/29 16:14:07 UTC
svn commit: r1635171 [2/2] - in /hive/trunk: itests/src/test/resources/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/java/org/apache/hadoop/hive/ql/exec/tez/
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/
ql/src/java/org/apache/hadoop/hive/ql...
Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java Wed Oct 29 15:14:06 2014
@@ -247,16 +247,16 @@ public class TestVectorizationContext {
children5.add(col6Expr);
modExpr.setChildren(children5);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
- columnMap.put("col3", 3);
- columnMap.put("col4", 4);
- columnMap.put("col5", 5);
- columnMap.put("col6", 6);
+ VectorizationContext vc = new VectorizationContext();
+ vc.addInitialColumn("col1");
+ vc.addInitialColumn("col2");
+ vc.addInitialColumn("col3");
+ vc.addInitialColumn("col4");
+ vc.addInitialColumn("col5");
+ vc.addInitialColumn("col6");
+ vc.finishedAddingInitialColumns();
//Generate vectorized expression
- VectorizationContext vc = new VectorizationContext(columnMap, 6);
VectorExpression ve = vc.getVectorExpression(sumExpr, VectorExpressionDescriptor.Mode.PROJECTION);
//Verify vectorized expression
@@ -293,11 +293,11 @@ public class TestVectorizationContext {
children1.add(constDesc);
exprDesc.setChildren(children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -318,11 +318,11 @@ public class TestVectorizationContext {
children1.add(col2Expr);
exprDesc.setChildren(children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -341,7 +341,7 @@ public class TestVectorizationContext {
children1.add(col2Expr);
exprDesc.setChildren(children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -360,7 +360,7 @@ public class TestVectorizationContext {
children1.add(col2Expr);
exprDesc.setChildren(children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -378,7 +378,7 @@ public class TestVectorizationContext {
children1.add(col2Expr);
exprDesc.setChildren(children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -395,7 +395,7 @@ public class TestVectorizationContext {
children1.add(col2Expr);
exprDesc.setChildren(children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -412,7 +412,7 @@ public class TestVectorizationContext {
children1.add(col2Expr);
exprDesc.setChildren(children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -432,10 +432,9 @@ public class TestVectorizationContext {
ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.doubleTypeInfo, udf,
children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col1");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
@@ -477,11 +476,11 @@ public class TestVectorizationContext {
children3.add(lessExprDesc);
andExprDesc.setChildren(children3);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
- columnMap.put("col2", 1);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(andExprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -528,11 +527,10 @@ public class TestVectorizationContext {
children3.add(col2Expr);
andExprDesc.setChildren(children3);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
- columnMap.put("col2", 1);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression veAnd = vc.getVectorExpression(andExprDesc, VectorExpressionDescriptor.Mode.FILTER);
assertEquals(veAnd.getClass(), FilterExprAndExpr.class);
assertEquals(veAnd.getChildExpressions()[0].getClass(), FilterLongColGreaterLongScalar.class);
@@ -557,7 +555,7 @@ public class TestVectorizationContext {
orExprDesc.setChildren(children4);
//Allocate new Vectorization context to reset the intermediate columns.
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
VectorExpression veOr = vc.getVectorExpression(orExprDesc, VectorExpressionDescriptor.Mode.FILTER);
assertEquals(veOr.getClass(), FilterExprOrExpr.class);
assertEquals(veOr.getChildExpressions()[0].getClass(), FilterLongColGreaterLongScalar.class);
@@ -594,11 +592,11 @@ public class TestVectorizationContext {
childOfNot.add(greaterExprDesc);
notExpr.setChildren(childOfNot);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
- columnMap.put("col2", 1);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(notExpr, VectorExpressionDescriptor.Mode.FILTER);
@@ -632,11 +630,10 @@ public class TestVectorizationContext {
childOfIsNull.add(greaterExprDesc);
isNullExpr.setChildren(childOfIsNull);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
- columnMap.put("col2", 1);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(isNullExpr, VectorExpressionDescriptor.Mode.FILTER);
@@ -674,11 +671,10 @@ public class TestVectorizationContext {
childOfNot.add(greaterExprDesc);
isNotNullExpr.setChildren(childOfNot);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
- columnMap.put("col2", 1);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(isNotNullExpr, VectorExpressionDescriptor.Mode.FILTER);
@@ -705,10 +701,9 @@ public class TestVectorizationContext {
ExprNodeGenericFuncDesc scalarMinusConstant = new ExprNodeGenericFuncDesc(TypeInfoFactory.longTypeInfo,
gudf, children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("a", 0);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(scalarMinusConstant, VectorExpressionDescriptor.Mode.PROJECTION);
assertEquals(ve.getClass(), LongScalarSubtractLongColumn.class);
@@ -727,11 +722,11 @@ public class TestVectorizationContext {
children1.add(constDesc);
exprDesc.setChildren(children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
@@ -746,9 +741,10 @@ public class TestVectorizationContext {
children.add(col1Expr);
ExprNodeGenericFuncDesc negExprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.longTypeInfo, gudf,
children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- VectorizationContext vc = new VectorizationContext(columnMap, 1);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(negExprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
@@ -763,9 +759,10 @@ public class TestVectorizationContext {
children.add(col1Expr);
ExprNodeGenericFuncDesc negExprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.doubleTypeInfo, gudf,
children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- VectorizationContext vc = new VectorizationContext(columnMap, 1);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(negExprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
@@ -788,10 +785,9 @@ public class TestVectorizationContext {
scalarGreaterColExpr.setChildren(children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("a", 0);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(scalarGreaterColExpr, VectorExpressionDescriptor.Mode.FILTER);
assertEquals(FilterLongScalarGreaterLongColumn.class, ve.getClass());
}
@@ -812,10 +808,9 @@ public class TestVectorizationContext {
colEqualScalar.setChildren(children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("a", 0);
-
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(colEqualScalar, VectorExpressionDescriptor.Mode.FILTER);
assertEquals(FilterLongColEqualLongScalar.class, ve.getClass());
}
@@ -836,9 +831,9 @@ public class TestVectorizationContext {
colEqualScalar.setChildren(children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("a", 0);
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(colEqualScalar, VectorExpressionDescriptor.Mode.PROJECTION);
assertEquals(LongColEqualLongScalar.class, ve.getClass());
}
@@ -852,11 +847,10 @@ public class TestVectorizationContext {
children.add(colDesc);
stringUnary.setChildren(children);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("b", 0);
- columnMap.put("a", 1);
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
-
+ List<String> columns = new ArrayList<String>();
+ columns.add("b");
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
GenericUDF stringLower = new GenericUDFLower();
stringUnary.setGenericUDF(stringLower);
@@ -866,7 +860,7 @@ public class TestVectorizationContext {
assertEquals(1, ((StringLower) ve).getColNum());
assertEquals(2, ((StringLower) ve).getOutputColumn());
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ExprNodeGenericFuncDesc anotherUnary = new ExprNodeGenericFuncDesc();
anotherUnary.setTypeInfo(TypeInfoFactory.stringTypeInfo);
@@ -898,10 +892,10 @@ public class TestVectorizationContext {
children1.add(colDesc1);
children2.add(colDesc2);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("b", 0);
- columnMap.put("a", 1);
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("b");
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
// Sin(double)
GenericUDFBridge gudfBridge = new GenericUDFBridge("sin", false, UDFSin.class.getName());
@@ -989,10 +983,10 @@ public class TestVectorizationContext {
List<ExprNodeDesc> children = new ArrayList<ExprNodeDesc>();
children.add(colDesc1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("b", 0);
- columnMap.put("a", 1);
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("b");
+ columns.add("a");
+ VectorizationContext vc = new VectorizationContext(columns);
//UDFYear
GenericUDFBridge gudfBridge = new GenericUDFBridge("year", false, UDFYear.class.getName());
@@ -1026,10 +1020,12 @@ public class TestVectorizationContext {
ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
+
VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
assertTrue(ve instanceof FilterStringColumnBetween);
@@ -1054,7 +1050,7 @@ public class TestVectorizationContext {
exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
assertTrue(ve instanceof FilterCharColumnBetween);
@@ -1079,7 +1075,7 @@ public class TestVectorizationContext {
exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
children1);
- vc = new VectorizationContext(columnMap, 2);
+ vc = new VectorizationContext(columns);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
assertTrue(ve instanceof FilterVarCharColumnBetween);
@@ -1144,10 +1140,11 @@ public class TestVectorizationContext {
ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo,
udf, children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
- VectorizationContext vc = new VectorizationContext(columnMap, 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
assertTrue(ve instanceof FilterStringColumnInList);
ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
@@ -1197,11 +1194,12 @@ public class TestVectorizationContext {
ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
children1);
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 1);
- columnMap.put("col2", 2);
- columnMap.put("col3", 3);
- VectorizationContext vc = new VectorizationContext(columnMap, 3);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ columns.add("col3");
+ VectorizationContext vc = new VectorizationContext(columns);
VectorExpression ve = vc.getVectorExpression(exprDesc);
assertTrue(ve instanceof IfExprLongColumnLongColumn);
Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java Wed Oct 29 15:14:06 2014
@@ -1298,8 +1298,8 @@ public class TestInputOutputFormat {
}
mapWork.setPathToAliases(aliasMap);
mapWork.setPathToPartitionInfo(partMap);
- mapWork.setScratchColumnMap(new HashMap<String, Map<String, Integer>>());
- mapWork.setScratchColumnVectorTypes(new HashMap<String,
+ mapWork.setAllColumnVectorMaps(new HashMap<String, Map<String, Integer>>());
+ mapWork.setAllScratchColumnVectorTypeMaps(new HashMap<String,
Map<Integer, String>>());
// write the plan out
Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java Wed Oct 29 15:14:06 2014
@@ -45,13 +45,14 @@ public class TestVectorizer {
@Before
public void setUp() {
- Map<String, Integer> columnMap = new HashMap<String, Integer>();
- columnMap.put("col1", 0);
- columnMap.put("col2", 1);
- columnMap.put("col3", 2);
+ List<String> columns = new ArrayList<String>();
+ columns.add("col0");
+ columns.add("col1");
+ columns.add("col2");
+ columns.add("col3");
//Generate vectorized expression
- vContext = new VectorizationContext(columnMap, 3);
+ vContext = new VectorizationContext(columns);
}
@Description(name = "fake", value = "FAKE")
Added: hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q?rev=1635171&view=auto
==============================================================================
--- hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q (added)
+++ hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q Wed Oct 29 15:14:06 2014
@@ -0,0 +1,11 @@
+SET hive.vectorized.execution.enabled=true;
+set hive.support.concurrency=true;
+set hive.enforce.bucketing=true;
+
+CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile;
+
+
+explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three'); select a, b from non_orc_table order by a;
+
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three'); select a, b from non_orc_table order by a;
Added: hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out?rev=1635171&view=auto
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out (added)
+++ hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out Wed Oct 29 15:14:06 2014
@@ -0,0 +1,105 @@
+PREHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@non_orc_table
+PREHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
+ Stage-0 depends on stages: Stage-2
+ Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Tez
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: values__tmp__table__1
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: tmp_values_col1 (type: string), tmp_values_col2 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ Reduce Output Operator
+ sort order:
+ Map-reduce partition columns: UDFToInteger(_col0) (type: int)
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ value expressions: _col0 (type: string), _col1 (type: string)
+ Reducer 2
+ Reduce Operator Tree:
+ Extract
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: UDFToInteger(_col0) (type: int), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.non_orc_table
+ Execution mode: vectorized
+
+ Stage: Stage-2
+ Dependency Collection
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.non_orc_table
+
+ Stage: Stage-3
+ Stats-Aggr Operator
+
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+PREHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+PREHOOK: Input: default@values__tmp__table__2
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@values__tmp__table__2
+POSTHOOK: Output: default@non_orc_table
+POSTHOOK: Lineage: non_orc_table.a EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: non_orc_table.b SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+1 one
+1 one
+2 two
+3 three
Modified: hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out Wed Oct 29 15:14:06 2014
@@ -110,6 +110,7 @@ STAGE PLANS:
sort order: +
Statistics: Num rows: 250 Data size: 49500 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: bigint), _col2 (type: bigint)
+ Execution mode: vectorized
Reducer 3
Reduce Operator Tree:
Select Operator
@@ -241,6 +242,7 @@ STAGE PLANS:
sort order: -
Statistics: Num rows: 250 Data size: 49500 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: bigint), _col2 (type: bigint)
+ Execution mode: vectorized
Reducer 3
Reduce Operator Tree:
Select Operator
Added: hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out?rev=1635171&view=auto
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out (added)
+++ hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out Wed Oct 29 15:14:06 2014
@@ -0,0 +1,94 @@
+PREHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@non_orc_table
+PREHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 depends on stages: Stage-1
+ Stage-2 depends on stages: Stage-0
+
+STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ alias: values__tmp__table__1
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: tmp_values_col1 (type: string), tmp_values_col2 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ Reduce Output Operator
+ sort order:
+ Map-reduce partition columns: UDFToInteger(_col0) (type: int)
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ value expressions: _col0 (type: string), _col1 (type: string)
+ Reduce Operator Tree:
+ Extract
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: UDFToInteger(_col0) (type: int), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.non_orc_table
+
+ Stage: Stage-0
+ Move Operator
+ tables:
+ replace: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.non_orc_table
+
+ Stage: Stage-2
+ Stats-Aggr Operator
+
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+PREHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+PREHOOK: Input: default@values__tmp__table__2
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@values__tmp__table__2
+POSTHOOK: Output: default@non_orc_table
+POSTHOOK: Lineage: non_orc_table.a EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: non_orc_table.b SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+1 one
+1 one
+2 two
+3 three