You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2014/10/29 16:14:07 UTC

svn commit: r1635171 [2/2] - in /hive/trunk: itests/src/test/resources/ ql/src/java/org/apache/hadoop/hive/ql/exec/ ql/src/java/org/apache/hadoop/hive/ql/exec/tez/ ql/src/java/org/apache/hadoop/hive/ql/exec/vector/ ql/src/java/org/apache/hadoop/hive/ql...

Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/vector/TestVectorizationContext.java Wed Oct 29 15:14:06 2014
@@ -247,16 +247,16 @@ public class TestVectorizationContext {
     children5.add(col6Expr);
     modExpr.setChildren(children5);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-    columnMap.put("col3", 3);
-    columnMap.put("col4", 4);
-    columnMap.put("col5", 5);
-    columnMap.put("col6", 6);
+    VectorizationContext vc = new VectorizationContext();
+    vc.addInitialColumn("col1");
+    vc.addInitialColumn("col2");
+    vc.addInitialColumn("col3");
+    vc.addInitialColumn("col4");
+    vc.addInitialColumn("col5");
+    vc.addInitialColumn("col6");
+    vc.finishedAddingInitialColumns();
 
     //Generate vectorized expression
-    VectorizationContext vc = new VectorizationContext(columnMap, 6);
     VectorExpression ve = vc.getVectorExpression(sumExpr, VectorExpressionDescriptor.Mode.PROJECTION);
 
     //Verify vectorized expression
@@ -293,11 +293,11 @@ public class TestVectorizationContext {
     children1.add(constDesc);
     exprDesc.setChildren(children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -318,11 +318,11 @@ public class TestVectorizationContext {
     children1.add(col2Expr);
     exprDesc.setChildren(children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -341,7 +341,7 @@ public class TestVectorizationContext {
     children1.add(col2Expr);
     exprDesc.setChildren(children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
 
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -360,7 +360,7 @@ public class TestVectorizationContext {
     children1.add(col2Expr);
     exprDesc.setChildren(children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
 
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -378,7 +378,7 @@ public class TestVectorizationContext {
     children1.add(col2Expr);
     exprDesc.setChildren(children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
 
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -395,7 +395,7 @@ public class TestVectorizationContext {
     children1.add(col2Expr);
     exprDesc.setChildren(children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
 
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -412,7 +412,7 @@ public class TestVectorizationContext {
     children1.add(col2Expr);
     exprDesc.setChildren(children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
 
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -432,10 +432,9 @@ public class TestVectorizationContext {
     ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.doubleTypeInfo, udf,
         children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col1");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
 
@@ -477,11 +476,11 @@ public class TestVectorizationContext {
     children3.add(lessExprDesc);
     andExprDesc.setChildren(children3);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-    columnMap.put("col2", 1);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(andExprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -528,11 +527,10 @@ public class TestVectorizationContext {
     children3.add(col2Expr);
     andExprDesc.setChildren(children3);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-    columnMap.put("col2", 1);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression veAnd = vc.getVectorExpression(andExprDesc, VectorExpressionDescriptor.Mode.FILTER);
     assertEquals(veAnd.getClass(), FilterExprAndExpr.class);
     assertEquals(veAnd.getChildExpressions()[0].getClass(), FilterLongColGreaterLongScalar.class);
@@ -557,7 +555,7 @@ public class TestVectorizationContext {
     orExprDesc.setChildren(children4);
 
     //Allocate new Vectorization context to reset the intermediate columns.
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
     VectorExpression veOr = vc.getVectorExpression(orExprDesc, VectorExpressionDescriptor.Mode.FILTER);
     assertEquals(veOr.getClass(), FilterExprOrExpr.class);
     assertEquals(veOr.getChildExpressions()[0].getClass(), FilterLongColGreaterLongScalar.class);
@@ -594,11 +592,11 @@ public class TestVectorizationContext {
     childOfNot.add(greaterExprDesc);
     notExpr.setChildren(childOfNot);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-    columnMap.put("col2", 1);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(notExpr, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -632,11 +630,10 @@ public class TestVectorizationContext {
     childOfIsNull.add(greaterExprDesc);
     isNullExpr.setChildren(childOfIsNull);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-    columnMap.put("col2", 1);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(isNullExpr, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -674,11 +671,10 @@ public class TestVectorizationContext {
     childOfNot.add(greaterExprDesc);
     isNotNullExpr.setChildren(childOfNot);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-    columnMap.put("col2", 1);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(isNotNullExpr, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -705,10 +701,9 @@ public class TestVectorizationContext {
     ExprNodeGenericFuncDesc scalarMinusConstant = new ExprNodeGenericFuncDesc(TypeInfoFactory.longTypeInfo,
         gudf, children);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("a", 0);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression ve = vc.getVectorExpression(scalarMinusConstant, VectorExpressionDescriptor.Mode.PROJECTION);
 
     assertEquals(ve.getClass(), LongScalarSubtractLongColumn.class);
@@ -727,11 +722,11 @@ public class TestVectorizationContext {
     children1.add(constDesc);
     exprDesc.setChildren(children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
 
@@ -746,9 +741,10 @@ public class TestVectorizationContext {
     children.add(col1Expr);
     ExprNodeGenericFuncDesc negExprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.longTypeInfo, gudf,
         children);
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    VectorizationContext vc = new VectorizationContext(columnMap, 1);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(negExprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
 
@@ -763,9 +759,10 @@ public class TestVectorizationContext {
     children.add(col1Expr);
     ExprNodeGenericFuncDesc negExprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.doubleTypeInfo, gudf,
         children);
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    VectorizationContext vc = new VectorizationContext(columnMap, 1);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     VectorExpression ve = vc.getVectorExpression(negExprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
 
@@ -788,10 +785,9 @@ public class TestVectorizationContext {
 
     scalarGreaterColExpr.setChildren(children);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("a", 0);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression ve = vc.getVectorExpression(scalarGreaterColExpr, VectorExpressionDescriptor.Mode.FILTER);
     assertEquals(FilterLongScalarGreaterLongColumn.class, ve.getClass());
   }
@@ -812,10 +808,9 @@ public class TestVectorizationContext {
 
     colEqualScalar.setChildren(children);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("a", 0);
-
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression ve = vc.getVectorExpression(colEqualScalar, VectorExpressionDescriptor.Mode.FILTER);
     assertEquals(FilterLongColEqualLongScalar.class, ve.getClass());
   }
@@ -836,9 +831,9 @@ public class TestVectorizationContext {
 
     colEqualScalar.setChildren(children);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("a", 0);
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression ve = vc.getVectorExpression(colEqualScalar, VectorExpressionDescriptor.Mode.PROJECTION);
     assertEquals(LongColEqualLongScalar.class, ve.getClass());
   }
@@ -852,11 +847,10 @@ public class TestVectorizationContext {
     children.add(colDesc);
     stringUnary.setChildren(children);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("b", 0);
-    columnMap.put("a", 1);
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
-
+    List<String> columns = new ArrayList<String>();
+    columns.add("b");
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
     GenericUDF stringLower = new GenericUDFLower();
     stringUnary.setGenericUDF(stringLower);
 
@@ -866,7 +860,7 @@ public class TestVectorizationContext {
     assertEquals(1, ((StringLower) ve).getColNum());
     assertEquals(2, ((StringLower) ve).getOutputColumn());
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
 
     ExprNodeGenericFuncDesc anotherUnary = new ExprNodeGenericFuncDesc();
     anotherUnary.setTypeInfo(TypeInfoFactory.stringTypeInfo);
@@ -898,10 +892,10 @@ public class TestVectorizationContext {
     children1.add(colDesc1);
     children2.add(colDesc2);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("b", 0);
-    columnMap.put("a", 1);
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("b");
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     // Sin(double)
     GenericUDFBridge gudfBridge = new GenericUDFBridge("sin", false, UDFSin.class.getName());
@@ -989,10 +983,10 @@ public class TestVectorizationContext {
     List<ExprNodeDesc> children = new ArrayList<ExprNodeDesc>();
     children.add(colDesc1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("b", 0);
-    columnMap.put("a", 1);
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("b");
+    columns.add("a");
+    VectorizationContext vc = new VectorizationContext(columns);
 
     //UDFYear
     GenericUDFBridge gudfBridge = new GenericUDFBridge("year", false, UDFYear.class.getName());
@@ -1026,10 +1020,12 @@ public class TestVectorizationContext {
     ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
         children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
+
     VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
     assertTrue(ve instanceof FilterStringColumnBetween);
 
@@ -1054,7 +1050,7 @@ public class TestVectorizationContext {
     exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
         children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
     assertTrue(ve instanceof FilterCharColumnBetween);
 
@@ -1079,7 +1075,7 @@ public class TestVectorizationContext {
     exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
         children1);
 
-    vc = new VectorizationContext(columnMap, 2);
+    vc = new VectorizationContext(columns);
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
     assertTrue(ve instanceof FilterVarCharColumnBetween);
 
@@ -1144,10 +1140,11 @@ public class TestVectorizationContext {
     ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo,
         udf, children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-    VectorizationContext vc = new VectorizationContext(columnMap, 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.FILTER);
     assertTrue(ve instanceof FilterStringColumnInList);
     ve = vc.getVectorExpression(exprDesc, VectorExpressionDescriptor.Mode.PROJECTION);
@@ -1197,11 +1194,12 @@ public class TestVectorizationContext {
     ExprNodeGenericFuncDesc exprDesc = new ExprNodeGenericFuncDesc(TypeInfoFactory.booleanTypeInfo, udf,
         children1);
 
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 1);
-    columnMap.put("col2", 2);
-    columnMap.put("col3", 3);
-    VectorizationContext vc = new VectorizationContext(columnMap, 3);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    columns.add("col3");
+    VectorizationContext vc = new VectorizationContext(columns);
     VectorExpression ve = vc.getVectorExpression(exprDesc);
     assertTrue(ve instanceof IfExprLongColumnLongColumn);
 

Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java Wed Oct 29 15:14:06 2014
@@ -1298,8 +1298,8 @@ public class TestInputOutputFormat {
     }
     mapWork.setPathToAliases(aliasMap);
     mapWork.setPathToPartitionInfo(partMap);
-    mapWork.setScratchColumnMap(new HashMap<String, Map<String, Integer>>());
-    mapWork.setScratchColumnVectorTypes(new HashMap<String,
+    mapWork.setAllColumnVectorMaps(new HashMap<String, Map<String, Integer>>());
+    mapWork.setAllScratchColumnVectorTypeMaps(new HashMap<String,
         Map<Integer, String>>());
 
     // write the plan out

Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/physical/TestVectorizer.java Wed Oct 29 15:14:06 2014
@@ -45,13 +45,14 @@ public class TestVectorizer {
 
   @Before
   public void setUp() {
-    Map<String, Integer> columnMap = new HashMap<String, Integer>();
-    columnMap.put("col1", 0);
-    columnMap.put("col2", 1);
-    columnMap.put("col3", 2);
+    List<String> columns = new ArrayList<String>();
+    columns.add("col0");
+    columns.add("col1");
+    columns.add("col2");
+    columns.add("col3");
 
     //Generate vectorized expression
-    vContext = new VectorizationContext(columnMap, 3);
+    vContext = new VectorizationContext(columns);
   }
 
   @Description(name = "fake", value = "FAKE")

Added: hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q?rev=1635171&view=auto
==============================================================================
--- hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q (added)
+++ hive/trunk/ql/src/test/queries/clientpositive/vector_bucket.q Wed Oct 29 15:14:06 2014
@@ -0,0 +1,11 @@
+SET hive.vectorized.execution.enabled=true;
+set hive.support.concurrency=true;
+set hive.enforce.bucketing=true;
+
+CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile; 
+
+
+explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three'); select a, b from non_orc_table order by a;
+
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three'); select a, b from non_orc_table order by a;

Added: hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out?rev=1635171&view=auto
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out (added)
+++ hive/trunk/ql/src/test/results/clientpositive/tez/vector_bucket.q.out Wed Oct 29 15:14:06 2014
@@ -0,0 +1,105 @@
+PREHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@non_orc_table
+PREHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-2 depends on stages: Stage-1
+  Stage-0 depends on stages: Stage-2
+  Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+  Stage: Stage-1
+    Tez
+      Edges:
+        Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: values__tmp__table__1
+                  Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                  Select Operator
+                    expressions: tmp_values_col1 (type: string), tmp_values_col2 (type: string)
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                    Reduce Output Operator
+                      sort order: 
+                      Map-reduce partition columns: UDFToInteger(_col0) (type: int)
+                      Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                      value expressions: _col0 (type: string), _col1 (type: string)
+        Reducer 2 
+            Reduce Operator Tree:
+              Extract
+                Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                Select Operator
+                  expressions: UDFToInteger(_col0) (type: int), _col1 (type: string)
+                  outputColumnNames: _col0, _col1
+                  Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.non_orc_table
+            Execution mode: vectorized
+
+  Stage: Stage-2
+    Dependency Collection
+
+  Stage: Stage-0
+    Move Operator
+      tables:
+          replace: false
+          table:
+              input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+              output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              name: default.non_orc_table
+
+  Stage: Stage-3
+    Stats-Aggr Operator
+
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+PREHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+PREHOOK: Input: default@values__tmp__table__2
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@values__tmp__table__2
+POSTHOOK: Output: default@non_orc_table
+POSTHOOK: Lineage: non_orc_table.a EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: non_orc_table.b SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+1	one
+1	one
+2	two
+3	three

Modified: hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out?rev=1635171&r1=1635170&r2=1635171&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/tez/vector_char_2.q.out Wed Oct 29 15:14:06 2014
@@ -110,6 +110,7 @@ STAGE PLANS:
                     sort order: +
                     Statistics: Num rows: 250 Data size: 49500 Basic stats: COMPLETE Column stats: NONE
                     value expressions: _col1 (type: bigint), _col2 (type: bigint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Select Operator
@@ -241,6 +242,7 @@ STAGE PLANS:
                     sort order: -
                     Statistics: Num rows: 250 Data size: 49500 Basic stats: COMPLETE Column stats: NONE
                     value expressions: _col1 (type: bigint), _col2 (type: bigint)
+            Execution mode: vectorized
         Reducer 3 
             Reduce Operator Tree:
               Select Operator

Added: hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out?rev=1635171&view=auto
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out (added)
+++ hive/trunk/ql/src/test/results/clientpositive/vector_bucket.q.out Wed Oct 29 15:14:06 2014
@@ -0,0 +1,94 @@
+PREHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: CREATE TABLE non_orc_table(a INT, b STRING) CLUSTERED BY(a) INTO 2 BUCKETS STORED AS sequencefile
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@non_orc_table
+PREHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+  Stage-2 depends on stages: Stage-0
+
+STAGE PLANS:
+  Stage: Stage-1
+    Map Reduce
+      Map Operator Tree:
+          TableScan
+            alias: values__tmp__table__1
+            Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+            Select Operator
+              expressions: tmp_values_col1 (type: string), tmp_values_col2 (type: string)
+              outputColumnNames: _col0, _col1
+              Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+              Reduce Output Operator
+                sort order: 
+                Map-reduce partition columns: UDFToInteger(_col0) (type: int)
+                Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+                value expressions: _col0 (type: string), _col1 (type: string)
+      Reduce Operator Tree:
+        Extract
+          Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+          Select Operator
+            expressions: UDFToInteger(_col0) (type: int), _col1 (type: string)
+            outputColumnNames: _col0, _col1
+            Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+            File Output Operator
+              compressed: false
+              Statistics: Num rows: 0 Data size: 26 Basic stats: PARTIAL Column stats: NONE
+              table:
+                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.non_orc_table
+
+  Stage: Stage-0
+    Move Operator
+      tables:
+          replace: false
+          table:
+              input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+              output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+              name: default.non_orc_table
+
+  Stage: Stage-2
+    Stats-Aggr Operator
+
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+PREHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+PREHOOK: type: QUERY
+PREHOOK: Input: default@values__tmp__table__2
+PREHOOK: Output: default@non_orc_table
+POSTHOOK: query: insert into table non_orc_table values(1, 'one'),(1, 'one'), (2, 'two'),(3, 'three')
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@values__tmp__table__2
+POSTHOOK: Output: default@non_orc_table
+POSTHOOK: Lineage: non_orc_table.a EXPRESSION [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, type:string, comment:), ]
+POSTHOOK: Lineage: non_orc_table.b SIMPLE [(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, type:string, comment:), ]
+PREHOOK: query: select a, b from non_orc_table order by a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+POSTHOOK: query: select a, b from non_orc_table order by a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@non_orc_table
+#### A masked pattern was here ####
+1	one
+1	one
+2	two
+3	three