You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2013/07/31 00:22:46 UTC

svn commit: r1508669 [7/39] - in /hive/branches/vectorization: ./ common/src/java/org/apache/hadoop/hive/conf/ conf/ contrib/src/test/results/clientpositive/ data/files/ eclipse-templates/ hcatalog/build-support/ant/ hcatalog/core/src/main/java/org/apa...

Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java Tue Jul 30 22:22:35 2013
@@ -76,6 +76,7 @@ public class GenericUDAFVariance extends
     case DECIMAL:
       return new GenericUDAFVarianceEvaluator();
     case BOOLEAN:
+    case DATE:
     default:
       throw new UDFArgumentTypeException(0,
           "Only numeric or string type arguments are accepted but "

Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java Tue Jul 30 22:22:35 2013
@@ -59,6 +59,7 @@ public class GenericUDAFVarianceSample e
     case DECIMAL:
       return new GenericUDAFVarianceSampleEvaluator();
     case BOOLEAN:
+    case DATE:
     default:
       throw new UDFArgumentTypeException(0,
           "Only numeric or string type arguments are accepted but "

Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java Tue Jul 30 22:22:35 2013
@@ -26,22 +26,15 @@ import org.apache.hadoop.hive.ql.exec.De
 import org.apache.hadoop.hive.ql.exec.UDFArgumentTypeException;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
 import org.apache.hadoop.hive.ql.parse.SemanticException;
-import org.apache.hadoop.hive.serde2.io.DoubleWritable;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
 import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.StandardMapObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.StandardListObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.StructField;
-import org.apache.hadoop.hive.serde2.objectinspector.primitive.DoubleObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableDoubleObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils;
-import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
 import org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo;
+import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-import org.apache.hadoop.util.StringUtils;
 import org.apache.hadoop.io.Text;
 
 /**
@@ -69,7 +62,7 @@ public class GenericUDAFnGrams implement
       throw new UDFArgumentTypeException(parameters.length-1,
           "Please specify either three or four arguments.");
     }
-    
+
     // Validate the first parameter, which is the expression to compute over. This should be an
     // array of strings type, or an array of arrays of strings.
     PrimitiveTypeInfo pti;
@@ -99,7 +92,7 @@ public class GenericUDAFnGrams implement
     }
     if(pti.getPrimitiveCategory() != PrimitiveObjectInspector.PrimitiveCategory.STRING) {
       throw new UDFArgumentTypeException(0,
-          "Only array<string> or array<array<string>> is allowed, but " 
+          "Only array<string> or array<array<string>> is allowed, but "
           + parameters[0].getTypeName() + " was passed as parameter 1.");
     }
 
@@ -107,7 +100,7 @@ public class GenericUDAFnGrams implement
     if(parameters[1].getCategory() != ObjectInspector.Category.PRIMITIVE) {
       throw new UDFArgumentTypeException(1, "Only integers are accepted but "
           + parameters[1].getTypeName() + " was passed as parameter 2.");
-    } 
+    }
     switch(((PrimitiveTypeInfo) parameters[1]).getPrimitiveCategory()) {
     case BYTE:
     case SHORT:
@@ -125,7 +118,7 @@ public class GenericUDAFnGrams implement
     if(parameters[2].getCategory() != ObjectInspector.Category.PRIMITIVE) {
       throw new UDFArgumentTypeException(2, "Only integers are accepted but "
             + parameters[2].getTypeName() + " was passed as parameter 3.");
-    } 
+    }
     switch(((PrimitiveTypeInfo) parameters[2]).getPrimitiveCategory()) {
     case BYTE:
     case SHORT:
@@ -144,7 +137,7 @@ public class GenericUDAFnGrams implement
       if(parameters[3].getCategory() != ObjectInspector.Category.PRIMITIVE) {
         throw new UDFArgumentTypeException(3, "Only integers are accepted but "
             + parameters[3].getTypeName() + " was passed as parameter 4.");
-      } 
+      }
       switch(((PrimitiveTypeInfo) parameters[3]).getPrimitiveCategory()) {
       case BYTE:
       case SHORT:
@@ -174,7 +167,7 @@ public class GenericUDAFnGrams implement
     private PrimitiveObjectInspector kOI;
     private PrimitiveObjectInspector pOI;
 
-    // For PARTIAL2 and FINAL: ObjectInspectors for partial aggregations 
+    // For PARTIAL2 and FINAL: ObjectInspectors for partial aggregations
     private StandardListObjectInspector loi;
 
     @Override
@@ -209,8 +202,8 @@ public class GenericUDAFnGrams implement
       // Init output object inspectors.
       //
       // The return type for a partial aggregation is still a list of strings.
-      // 
-      // The return type for FINAL and COMPLETE is a full aggregation result, which is 
+      //
+      // The return type for FINAL and COMPLETE is a full aggregation result, which is
       // an array of structures containing the n-gram and its estimated frequency.
       if (m == Mode.PARTIAL1 || m == Mode.PARTIAL2) {
         return ObjectInspectorFactory.getStandardListObjectInspector(
@@ -224,7 +217,7 @@ public class GenericUDAFnGrams implement
         foi.add(PrimitiveObjectInspectorFactory.writableDoubleObjectInspector);
         ArrayList<String> fname = new ArrayList<String>();
         fname.add("ngram");
-        fname.add("estfrequency");               
+        fname.add("estfrequency");
         return ObjectInspectorFactory.getStandardListObjectInspector(
                  ObjectInspectorFactory.getStandardStructObjectInspector(fname, foi) );
       }
@@ -232,14 +225,14 @@ public class GenericUDAFnGrams implement
 
     @Override
     public void merge(AggregationBuffer agg, Object partial) throws HiveException {
-      if(partial == null) { 
+      if(partial == null) {
         return;
       }
       NGramAggBuf myagg = (NGramAggBuf) agg;
       List<Text> partialNGrams = (List<Text>) loi.getList(partial);
       int n = Integer.parseInt(partialNGrams.get(partialNGrams.size()-1).toString());
       if(myagg.n > 0 && myagg.n != n) {
-        throw new HiveException(getClass().getSimpleName() + ": mismatch in value for 'n'" 
+        throw new HiveException(getClass().getSimpleName() + ": mismatch in value for 'n'"
             + ", which usually is caused by a non-constant expression. Found '"+n+"' and '"
             + myagg.n + "'.");
       }
@@ -273,7 +266,7 @@ public class GenericUDAFnGrams implement
         return;
       }
       NGramAggBuf myagg = (NGramAggBuf) agg;
-    
+
       // Parse out 'n' and 'k' if we haven't already done so, and while we're at it,
       // also parse out the precision factor 'pf' if the user has supplied one.
       if(!myagg.nge.isInitialized()) {
@@ -317,7 +310,7 @@ public class GenericUDAFnGrams implement
 
           // parse out n-grams, update frequency counts
           processNgrams(myagg, words);
-        } 
+        }
       } else {
         // we're dealing with an array of strings
         ArrayList<String> words = new ArrayList<String>();
@@ -337,7 +330,7 @@ public class GenericUDAFnGrams implement
       return myagg.nge.getNGrams();
     }
 
-    // Aggregation buffer methods. 
+    // Aggregation buffer methods.
     static class NGramAggBuf extends AbstractAggregationBuffer {
       NGramEstimator nge;
       int n;

Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java Tue Jul 30 22:22:35 2013
@@ -136,10 +136,18 @@ public abstract class GenericUDFBaseComp
       if (oiTypeInfo0 != oiTypeInfo1) {
         compareType = CompareType.NEED_CONVERT;
 
-        // If either argument is a string, we convert to a double because a number
-        // in string form should always be convertible into a double
-        if (oiTypeInfo0.equals(TypeInfoFactory.stringTypeInfo)
+        if ((oiTypeInfo0.equals(TypeInfoFactory.stringTypeInfo)
+                && oiTypeInfo1.equals(TypeInfoFactory.dateTypeInfo))
+            || (oiTypeInfo0.equals(TypeInfoFactory.dateTypeInfo)
+                && oiTypeInfo1.equals(TypeInfoFactory.stringTypeInfo))) {
+          // Date should be comparable with string
+          compareOI = TypeInfoUtils.getStandardWritableObjectInspectorFromTypeInfo(
+              TypeInfoFactory.stringTypeInfo);
+
+        } else if (oiTypeInfo0.equals(TypeInfoFactory.stringTypeInfo)
             || oiTypeInfo1.equals(TypeInfoFactory.stringTypeInfo)) {
+          // If either argument is a string, we convert to a double because a number
+          // in string form should always be convertible into a double
           compareOI = TypeInfoUtils.getStandardWritableObjectInspectorFromTypeInfo(
               TypeInfoFactory.doubleTypeInfo);
         } else {

Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java Tue Jul 30 22:22:35 2013
@@ -28,6 +28,7 @@ import org.apache.hadoop.hive.ql.exec.UD
 import org.apache.hadoop.hive.ql.exec.UDFArgumentLengthException;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
+import org.apache.hadoop.hive.serde2.objectinspector.primitive.DateObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.StringObjectInspector;
 import org.apache.hadoop.hive.serde2.objectinspector.primitive.TimestampObjectInspector;
@@ -42,6 +43,7 @@ import org.apache.hadoop.io.LongWritable
 public class GenericUDFToUnixTimeStamp extends GenericUDF {
 
   private StringObjectInspector intputTextOI;
+  private DateObjectInspector inputDateOI;
   private TimestampObjectInspector inputTimestampOI;
   private StringObjectInspector patternOI;
 
@@ -69,6 +71,8 @@ public class GenericUDFToUnixTimeStamp e
         }
         patternOI = (StringObjectInspector) arguments[1];
       }
+    } else if (arguments[0] instanceof DateObjectInspector) {
+      inputDateOI = (DateObjectInspector) arguments[0];
     } else if (arguments[0] instanceof TimestampObjectInspector) {
       inputTimestampOI = (TimestampObjectInspector) arguments[0];
     } else {
@@ -106,7 +110,11 @@ public class GenericUDFToUnixTimeStamp e
       } catch (ParseException e) {
         return null;
       }
-    }
+    } else if (inputDateOI != null) {
+      retValue.set(inputDateOI.getPrimitiveWritableObject(arguments[0].get())
+                   .getTimeInSeconds());
+      return retValue;
+	}
     Timestamp timestamp = inputTimestampOI.getPrimitiveJavaObject(arguments[0].get());
     retValue.set(timestamp.getTime() / 1000);
     return retValue;

Modified: hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto (original)
+++ hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto Tue Jul 30 22:22:35 2013
@@ -27,6 +27,12 @@ message DecimalStatistics {
   optional string sum = 3;
 }
 
+message DateStatistics {
+  // min,max values saved as days since epoch
+  optional sint32 minimum = 1;
+  optional sint32 maximum = 2;
+}
+
 message ColumnStatistics {
   optional uint64 numberOfValues = 1;
   optional IntegerStatistics intStatistics = 2;
@@ -34,6 +40,7 @@ message ColumnStatistics {
   optional StringStatistics stringStatistics = 4;
   optional BucketStatistics bucketStatistics = 5;
   optional DecimalStatistics decimalStatistics = 6;
+  optional DateStatistics dateStatistics = 7;
 }
 
 message RowIndexEntry {
@@ -93,6 +100,7 @@ message Type {
     STRUCT = 12;
     UNION = 13;
     DECIMAL = 14;
+    DATE = 15;
   }
   required Kind kind = 1;
   repeated uint32 subtypes = 2 [packed=true];

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java Tue Jul 30 22:22:35 2013
@@ -886,7 +886,7 @@ public class QTestUtil {
 
       FileOutputStream ofs = new FileOutputStream(outf);
       for (Task<? extends Serializable> plan : tasks) {
-        Utilities.serializeTasks(plan, ofs);
+        Utilities.serializeObject(plan, ofs);
       }
 
       String[] patterns = new String[] {

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java Tue Jul 30 22:22:35 2013
@@ -52,6 +52,7 @@ import org.apache.hadoop.hive.ql.plan.Ma
 import org.apache.hadoop.hive.ql.plan.PartitionDesc;
 import org.apache.hadoop.hive.ql.plan.PlanUtils;
 import org.apache.hadoop.hive.ql.plan.ReduceSinkDesc;
+import org.apache.hadoop.hive.ql.plan.ReduceWork;
 import org.apache.hadoop.hive.ql.plan.ScriptDesc;
 import org.apache.hadoop.hive.ql.plan.SelectDesc;
 import org.apache.hadoop.hive.serde.serdeConstants;
@@ -141,7 +142,7 @@ public class TestExecDriver extends Test
   }
 
   public static void addMapWork(MapredWork mr, Table tbl, String alias, Operator<?> work) {
-    mr.addMapWork(tbl.getDataLocation().toString(), alias, work, new PartitionDesc(
+    mr.getMapWork().addMapWork(tbl.getDataLocation().toString(), alias, work, new PartitionDesc(
         Utilities.getTableDesc(tbl), null));
   }
 
@@ -194,7 +195,6 @@ public class TestExecDriver extends Test
 
   @SuppressWarnings("unchecked")
   private void populateMapPlan1(Table src) {
-    mr.setNumReduceTasks(Integer.valueOf(0));
 
     Operator<FileSinkDesc> op2 = OperatorFactory.get(new FileSinkDesc(tmpdir
         + "mapplan1.out", Utilities.defaultTd, true));
@@ -206,7 +206,6 @@ public class TestExecDriver extends Test
 
   @SuppressWarnings("unchecked")
   private void populateMapPlan2(Table src) {
-    mr.setNumReduceTasks(Integer.valueOf(0));
 
     Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
         + "mapplan2.out", Utilities.defaultTd, false));
@@ -225,7 +224,6 @@ public class TestExecDriver extends Test
 
   @SuppressWarnings("unchecked")
   private void populateMapRedPlan1(Table src) throws SemanticException {
-    mr.setNumReduceTasks(Integer.valueOf(1));
 
     ArrayList<String> outputColumns = new ArrayList<String>();
     for (int i = 0; i < 2; i++) {
@@ -238,8 +236,11 @@ public class TestExecDriver extends Test
         -1, 1, -1));
 
     addMapWork(mr, src, "a", op1);
-    mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
-    mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    ReduceWork rWork = new ReduceWork();
+    rWork.setNumReduceTasks(Integer.valueOf(1));
+    rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+    rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    mr.setReduceWork(rWork);
 
     // reduce side work
     Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -248,12 +249,11 @@ public class TestExecDriver extends Test
     Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
         getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
 
-    mr.setReducer(op2);
+    rWork.setReducer(op2);
   }
 
   @SuppressWarnings("unchecked")
   private void populateMapRedPlan2(Table src) throws SemanticException {
-    mr.setNumReduceTasks(Integer.valueOf(1));
     ArrayList<String> outputColumns = new ArrayList<String>();
     for (int i = 0; i < 2; i++) {
       outputColumns.add("_col" + i);
@@ -266,8 +266,11 @@ public class TestExecDriver extends Test
         outputColumns, false, -1, 1, -1));
 
     addMapWork(mr, src, "a", op1);
-    mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
-    mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    ReduceWork rWork = new ReduceWork();
+    rWork.setNumReduceTasks(Integer.valueOf(1));
+    rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+    rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    mr.setReduceWork(rWork);
 
     // reduce side work
     Operator<FileSinkDesc> op4 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -278,7 +281,7 @@ public class TestExecDriver extends Test
     Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
         getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
 
-    mr.setReducer(op2);
+    rWork.setReducer(op2);
   }
 
   /**
@@ -286,8 +289,6 @@ public class TestExecDriver extends Test
    */
   @SuppressWarnings("unchecked")
   private void populateMapRedPlan3(Table src, Table src2) throws SemanticException {
-    mr.setNumReduceTasks(Integer.valueOf(5));
-    mr.setNeedsTagging(true);
     List<String> outputColumns = new ArrayList<String>();
     for (int i = 0; i < 2; i++) {
       outputColumns.add("_col" + i);
@@ -299,8 +300,6 @@ public class TestExecDriver extends Test
         Byte.valueOf((byte) 0), 1, -1));
 
     addMapWork(mr, src, "a", op1);
-    mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
-    mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
 
     Operator<ReduceSinkDesc> op2 = OperatorFactory.get(PlanUtils
         .getReduceSinkDesc(Utilities.makeList(getStringColumn("key")),
@@ -308,7 +307,14 @@ public class TestExecDriver extends Test
         Byte.valueOf((byte) 1), Integer.MAX_VALUE, -1));
 
     addMapWork(mr, src2, "b", op2);
-    mr.getTagToValueDesc().add(op2.getConf().getValueSerializeInfo());
+    ReduceWork rWork = new ReduceWork();
+    rWork.setNumReduceTasks(Integer.valueOf(5));
+    rWork.setNeedsTagging(true);
+    rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+    rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+
+    mr.setReduceWork(rWork);
+    rWork.getTagToValueDesc().add(op2.getConf().getValueSerializeInfo());
 
     // reduce side work
     Operator<FileSinkDesc> op4 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -320,12 +326,11 @@ public class TestExecDriver extends Test
         Utilities.ReduceField.VALUE.toString(), "", false), "0", false)),
         Utilities.makeList(outputColumns.get(0))), op4);
 
-    mr.setReducer(op5);
+    rWork.setReducer(op5);
   }
 
   @SuppressWarnings("unchecked")
   private void populateMapRedPlan4(Table src) throws SemanticException {
-    mr.setNumReduceTasks(Integer.valueOf(1));
 
     // map-side work
     ArrayList<String> outputColumns = new ArrayList<String>();
@@ -348,8 +353,11 @@ public class TestExecDriver extends Test
         outputColumns), op0);
 
     addMapWork(mr, src, "a", op4);
-    mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
-    mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    ReduceWork rWork = new ReduceWork();
+    rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+    rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    rWork.setNumReduceTasks(Integer.valueOf(1));
+    mr.setReduceWork(rWork);
 
     // reduce side work
     Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -358,7 +366,7 @@ public class TestExecDriver extends Test
     Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
         getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
 
-    mr.setReducer(op2);
+    rWork.setReducer(op2);
   }
 
   public static ExprNodeColumnDesc getStringColumn(String columnName) {
@@ -368,7 +376,6 @@ public class TestExecDriver extends Test
 
   @SuppressWarnings("unchecked")
   private void populateMapRedPlan5(Table src) throws SemanticException {
-    mr.setNumReduceTasks(Integer.valueOf(1));
 
     // map-side work
     ArrayList<String> outputColumns = new ArrayList<String>();
@@ -385,8 +392,11 @@ public class TestExecDriver extends Test
         outputColumns), op0);
 
     addMapWork(mr, src, "a", op4);
-    mr.setKeyDesc(op0.getConf().getKeySerializeInfo());
-    mr.getTagToValueDesc().add(op0.getConf().getValueSerializeInfo());
+    ReduceWork rWork = new ReduceWork();
+    mr.setReduceWork(rWork);
+    rWork.setNumReduceTasks(Integer.valueOf(1));
+    rWork.setKeyDesc(op0.getConf().getKeySerializeInfo());
+    rWork.getTagToValueDesc().add(op0.getConf().getValueSerializeInfo());
 
     // reduce side work
     Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -395,12 +405,11 @@ public class TestExecDriver extends Test
     Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
         getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
 
-    mr.setReducer(op2);
+    rWork.setReducer(op2);
   }
 
   @SuppressWarnings("unchecked")
   private void populateMapRedPlan6(Table src) throws SemanticException {
-    mr.setNumReduceTasks(Integer.valueOf(1));
 
     // map-side work
     ArrayList<String> outputColumns = new ArrayList<String>();
@@ -424,8 +433,11 @@ public class TestExecDriver extends Test
         outputColumns), op0);
 
     addMapWork(mr, src, "a", op4);
-    mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
-    mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+    ReduceWork rWork = new ReduceWork();
+    mr.setReduceWork(rWork);
+    rWork.setNumReduceTasks(Integer.valueOf(1));
+    rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+    rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
 
     // reduce side work
     Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -436,7 +448,7 @@ public class TestExecDriver extends Test
     Operator<ExtractDesc> op5 = OperatorFactory.get(new ExtractDesc(
         getStringColumn(Utilities.ReduceField.VALUE.toString())), op2);
 
-    mr.setReducer(op5);
+    rWork.setReducer(op5);
   }
 
   private void executePlan() throws Exception {

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java Tue Jul 30 22:22:35 2013
@@ -18,20 +18,20 @@
 
 package org.apache.hadoop.hive.ql.exec;
 
-import java.util.List;
-import java.util.LinkedList;
 import java.lang.reflect.Method;
+import java.util.LinkedList;
+import java.util.List;
 
 import junit.framework.TestCase;
 
-import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-import org.apache.hadoop.hive.serde.serdeConstants;
-import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
 import org.apache.hadoop.hive.serde2.io.DoubleWritable;
 import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable;
-import org.apache.hadoop.io.IntWritable;
-import org.apache.hadoop.io.BytesWritable;
 import org.apache.hadoop.hive.serde2.io.TimestampWritable;
+import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
+import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
+import org.apache.hadoop.io.BytesWritable;
+import org.apache.hadoop.io.IntWritable;
 
 public class TestFunctionRegistry extends TestCase {
 
@@ -41,10 +41,11 @@ public class TestFunctionRegistry extend
     public void one(IntWritable x, HiveDecimalWritable y) {}
     public void one(IntWritable x, DoubleWritable y) {}
     public void one(IntWritable x, IntWritable y) {}
+    public void mismatch(DateWritable x, HiveDecimalWritable y) {}
     public void mismatch(TimestampWritable x, HiveDecimalWritable y) {}
     public void mismatch(BytesWritable x, DoubleWritable y) {}
   }
-  
+
   @Override
   protected void setUp() {
   }
@@ -58,17 +59,18 @@ public class TestFunctionRegistry extend
     implicit(TypeInfoFactory.floatTypeInfo, TypeInfoFactory.decimalTypeInfo, true);
     implicit(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo, true);
     implicit(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo, true);
+    implicit(TypeInfoFactory.dateTypeInfo, TypeInfoFactory.decimalTypeInfo, false);
     implicit(TypeInfoFactory.timestampTypeInfo, TypeInfoFactory.decimalTypeInfo, false);
   }
 
-  private void verify(Class udf, String name, TypeInfo ta, TypeInfo tb, 
+  private void verify(Class udf, String name, TypeInfo ta, TypeInfo tb,
                       Class a, Class b, boolean throwException) {
     List<TypeInfo> args = new LinkedList<TypeInfo>();
     args.add(ta);
     args.add(tb);
 
     Method result = null;
-    
+
     try {
       result = FunctionRegistry.getMethodInternal(udf, name, false, args);
     } catch (UDFArgumentException e) {
@@ -116,13 +118,13 @@ public class TestFunctionRegistry extend
   }
 
   public void testCommonClass() {
-    common(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo, 
+    common(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo,
            TypeInfoFactory.decimalTypeInfo);
-    common(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo, 
+    common(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo,
            TypeInfoFactory.stringTypeInfo);
-    common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo, 
+    common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo,
            TypeInfoFactory.decimalTypeInfo);
-    common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo, 
+    common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo,
            TypeInfoFactory.stringTypeInfo);
   }
 
@@ -131,13 +133,13 @@ public class TestFunctionRegistry extend
   }
 
   public void testCommonClassComparison() {
-    comparison(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo, 
+    comparison(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo,
                TypeInfoFactory.decimalTypeInfo);
-    comparison(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo, 
+    comparison(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo,
                TypeInfoFactory.decimalTypeInfo);
-    comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo, 
+    comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo,
                TypeInfoFactory.decimalTypeInfo);
-    comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo, 
+    comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo,
                TypeInfoFactory.doubleTypeInfo);
   }
 

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java Tue Jul 30 22:22:35 2013
@@ -345,13 +345,13 @@ public class TestOperators extends TestC
 
       // initialize mapredWork
       MapredWork mrwork = new MapredWork();
-      mrwork.setPathToAliases(pathToAliases);
-      mrwork.setPathToPartitionInfo(pathToPartitionInfo);
-      mrwork.setAliasToWork(aliasToWork);
+      mrwork.getMapWork().setPathToAliases(pathToAliases);
+      mrwork.getMapWork().setPathToPartitionInfo(pathToPartitionInfo);
+      mrwork.getMapWork().setAliasToWork(aliasToWork);
 
       // get map operator and initialize it
       MapOperator mo = new MapOperator();
-      mo.initializeAsRoot(hconf, mrwork);
+      mo.initializeAsRoot(hconf, mrwork.getMapWork());
 
       Text tw = new Text();
       InspectableObject io1 = new InspectableObject();

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java Tue Jul 30 22:22:35 2013
@@ -75,13 +75,13 @@ public class TestPlan extends TestCase {
       ao.put("a", op);
 
       MapredWork mrwork = new MapredWork();
-      mrwork.setPathToAliases(pa);
-      mrwork.setPathToPartitionInfo(pt);
-      mrwork.setAliasToWork(ao);
+      mrwork.getMapWork().setPathToAliases(pa);
+      mrwork.getMapWork().setPathToPartitionInfo(pt);
+      mrwork.getMapWork().setAliasToWork(ao);
 
       // serialize the configuration once ..
       ByteArrayOutputStream baos = new ByteArrayOutputStream();
-      Utilities.serializeMapRedWork(mrwork, baos);
+      Utilities.serializeObject(mrwork, baos);
       baos.close();
       String v1 = baos.toString();
 
@@ -91,7 +91,7 @@ public class TestPlan extends TestCase {
       Utilities.setMapRedWork(job, mrwork, System.getProperty("java.io.tmpdir") + File.separator +
         System.getProperty("user.name") + File.separator + "hive");
       MapredWork mrwork2 = Utilities.getMapRedWork(job);
-      Utilities.clearMapRedWork(job);
+      Utilities.clearWork(job);
 
       // over here we should have some checks of the deserialized object against
       // the orginal object
@@ -99,7 +99,7 @@ public class TestPlan extends TestCase {
 
       // serialize again
       baos.reset();
-      Utilities.serializeMapRedWork(mrwork2, baos);
+      Utilities.serializeObject(mrwork2, baos);
       baos.close();
 
       // verify that the two are equal

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java Tue Jul 30 22:22:35 2013
@@ -18,11 +18,18 @@
 
 package org.apache.hadoop.hive.ql.exec;
 
+import java.sql.Date;
+import java.sql.Timestamp;
+
 import static org.apache.hadoop.hive.ql.exec.Utilities.getFileExtension;
 
 import junit.framework.TestCase;
 
+import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat;
+import org.apache.hadoop.hive.ql.plan.ExprNodeConstantDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
+import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
 import org.apache.hadoop.mapred.JobConf;
 
 public class TestUtilities extends TestCase {
@@ -53,4 +60,14 @@ public class TestUtilities extends TestC
     assertEquals("Custom extension for uncompressed text format", extension,
         getFileExtension(jc, true, new HiveIgnoreKeyTextOutputFormat()));
   }
+
+  public void testSerializeTimestamp() {
+    Timestamp ts = new Timestamp(1374554702000L);
+    ts.setNanos(123456);
+    ExprNodeConstantDesc constant = new ExprNodeConstantDesc(
+        TypeInfoFactory.timestampTypeInfo, ts);
+    String serialized = Utilities.serializeExpression(constant);
+    ExprNodeDesc deserialized = Utilities.deserializeExpression(serialized, new Configuration());
+    assertEquals(constant.getExprString(), deserialized.getExprString());
+  }
 }

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java Tue Jul 30 22:22:35 2013
@@ -38,7 +38,7 @@ public class VerifyHiveSortedInputFormat
       for (Task<? extends Serializable> rootTask : rootTasks) {
         if (rootTask.getWork() instanceof MapredWork) {
           Assert.assertTrue("The root map reduce task's input was not marked as sorted.",
-              ((MapredWork)rootTask.getWork()).isInputFormatSorted());
+              ((MapredWork)rootTask.getWork()).getMapWork().isInputFormatSorted());
         }
       }
     }

Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java Tue Jul 30 22:22:35 2013
@@ -46,7 +46,6 @@ import org.apache.hadoop.mapred.InputSpl
 import org.apache.hadoop.mapred.JobConf;
 import org.apache.hadoop.mapred.RecordReader;
 import org.apache.hadoop.mapred.Reporter;
-import org.apache.hadoop.mapred.TextInputFormat;
 import org.apache.hadoop.util.ReflectionUtils;
 
 /**
@@ -167,8 +166,8 @@ public class TestSymlinkTextInputFormat 
       
       QueryPlan plan = drv.getPlan();
       MapRedTask selectTask = (MapRedTask)plan.getRootTasks().get(0);
-      
-      ExecDriver.addInputPaths(newJob, selectTask.getWork(), emptyScratchDir.toString(), ctx);
+
+      ExecDriver.addInputPaths(newJob, selectTask.getWork().getMapWork(), emptyScratchDir.toString(), ctx);
       Utilities.setMapRedWork(newJob, selectTask.getWork(), ctx.getMRTmpFileURI());
       
       CombineHiveInputFormat combineInputFormat = ReflectionUtils.newInstance(

Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q Tue Jul 30 22:22:35 2013
@@ -1,2 +1,2 @@
 CREATE TABLE alter_test (d STRING);
-ALTER TABLE alter_test CHANGE d d DATE;
+ALTER TABLE alter_test CHANGE d d DATETIME;

Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q Tue Jul 30 22:22:35 2013
@@ -1,2 +1,2 @@
 CREATE TABLE alter_test (d STRING);
-ALTER TABLE alter_test ADD COLUMNS (ds DATE);
+ALTER TABLE alter_test ADD COLUMNS (ds DATETIME);

Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q Tue Jul 30 22:22:35 2013
@@ -1 +1 @@
-SELECT TRANSFORM(*) USING 'cat' AS (key DATE) FROM src;
+SELECT TRANSFORM(*) USING 'cat' AS (key DATETIME) FROM src;

Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q Tue Jul 30 22:22:35 2013
@@ -1,5 +1,5 @@
 USE default;
---  This should fail because Regex SerDe doesn't support TIMESTAMP, STRUCT
+--  This should fail because Regex SerDe doesn't support STRUCT
 CREATE TABLE serde_regex(
   host STRING,
   identity STRING,

Modified: hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q Tue Jul 30 22:22:35 2013
@@ -12,8 +12,8 @@ CREATE TABLE test_serializer(string1 STR
                              struct1 STRUCT<sint:INT,sboolean:BOOLEAN,sstring:STRING>,
                              enum1 STRING,
                              nullableint INT,
-                             bytes1 ARRAY<TINYINT>,
-                             fixed1 ARRAY<TINYINT>)
+                             bytes1 BINARY,
+                             fixed1 BINARY)
  ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' COLLECTION ITEMS TERMINATED BY ':' MAP KEYS TERMINATED BY '#' LINES TERMINATED BY '\n'
  STORED AS TEXTFILE;
 

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out Tue Jul 30 22:22:35 2013
@@ -45,7 +45,7 @@ SORT BY k1, v1, k2, v2 
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -72,7 +72,7 @@ SORT BY k1, v1, k2, v2 
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -205,7 +205,7 @@ SORT BY k1, v1, k2, v2 
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -232,7 +232,7 @@ SORT BY k1, v1, k2, v2 
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -281,7 +281,7 @@ SORT BY k1, v1, k2, v2 
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -303,7 +303,7 @@ SORT BY k1, v1, k2, v2 
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out Tue Jul 30 22:22:35 2013
@@ -51,7 +51,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -76,7 +76,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 80.0) and (key < 100.0))'
+'                  expr: ((key < 80) and (key < 100))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -103,7 +103,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -191,7 +191,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -214,7 +214,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 80.0) and (key < 100.0))'
+'                  expr: ((key < 80) and (key < 100))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -241,7 +241,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -302,7 +302,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -325,7 +325,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -354,7 +354,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 80.0) and (key < 100.0))'
+'                  expr: ((key < 80) and (key < 100))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -406,7 +406,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -430,7 +430,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 100.0) and (key < 80.0))'
+'                  expr: ((key < 100) and (key < 80))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -456,7 +456,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 80.0) and (key < 100.0))'
+'                  expr: ((key < 80) and (key < 100))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out Tue Jul 30 22:22:35 2013
@@ -50,7 +50,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 100.0)'
+'                  expr: (key < 100)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -77,7 +77,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 100.0)'
+'                  expr: (key < 100)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -120,7 +120,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -230,7 +230,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -294,7 +294,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -351,7 +351,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 100.0)'
+'                  expr: (key < 100)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -376,7 +376,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 100.0)'
+'                  expr: (key < 100)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -414,7 +414,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 100.0)'
+'                  expr: (key < 100)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -438,7 +438,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 100.0)'
+'                  expr: (key < 100)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out Tue Jul 30 22:22:35 2013
@@ -42,7 +42,7 @@ where tab.value < 200;
 '            alias: tab'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key > 20.0) and (value < 200.0))'
+'                  expr: ((key > 20) and (value < 200))'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -62,7 +62,7 @@ where tab.value < 200;
 '            alias: a'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (((key > 10.0) and (key > 20.0)) and (value < 200.0))'
+'                  expr: (((key > 10) and (key > 20)) and (value < 200))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -146,7 +146,7 @@ where tab.value < 200;
 '            alias: a'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (((key > 10.0) and (key > 20.0)) and (value < 200.0))'
+'                  expr: (((key > 10) and (key > 20)) and (value < 200))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -173,7 +173,7 @@ where tab.value < 200;
 '            alias: tab'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key > 20.0) and (value < 200.0))'
+'                  expr: ((key > 20) and (value < 200))'
 '                  type: boolean'
 '              Map Join Operator'
 '                condition map:'
@@ -217,7 +217,7 @@ where tab.value < 200;
 '            alias: a'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (((key > 10.0) and (key > 20.0)) and (value < 200.0))'
+'                  expr: (((key > 10) and (key > 20)) and (value < 200))'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -247,7 +247,7 @@ where tab.value < 200;
 '            alias: tab'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key > 20.0) and (value < 200.0))'
+'                  expr: ((key > 20) and (value < 200))'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out Tue Jul 30 22:22:35 2013
@@ -42,7 +42,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -52,7 +52,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '                filter predicates:'
 '                  0 '
 '                  1 '
-'                  2 {(key < 20.0)}'
+'                  2 {(key < 20)}'
 '                handleSkewJoin: false'
 '                keys:'
 '                  0 [Column[key]]'
@@ -64,7 +64,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -74,7 +74,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '                filter predicates:'
 '                  0 '
 '                  1 '
-'                  2 {(key < 20.0)}'
+'                  2 {(key < 20)}'
 '                handleSkewJoin: false'
 '                keys:'
 '                  0 [Column[key]]'
@@ -99,7 +99,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '              filter predicates:'
 '                0 '
 '                1 '
-'                2 {(key < 20.0)}'
+'                2 {(key < 20)}'
 '              handleSkewJoin: false'
 '              keys:'
 '                0 [Column[key]]'
@@ -231,7 +231,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -252,7 +252,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -297,7 +297,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '          filter predicates:'
 '            0 '
 '            1 '
-'            2 {(VALUE._col0 < 20.0)}'
+'            2 {(VALUE._col0 < 20)}'
 '          handleSkewJoin: false'
 '          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
 '          Select Operator'
@@ -378,7 +378,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 10.0) and (key < 15.0))'
+'                  expr: ((key < 10) and (key < 15))'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -388,7 +388,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '                filter predicates:'
 '                  0 '
 '                  1 '
-'                  2 {(key < 20.0)}'
+'                  2 {(key < 20)}'
 '                handleSkewJoin: false'
 '                keys:'
 '                  0 [Column[key]]'
@@ -400,7 +400,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 15.0) and (key < 10.0))'
+'                  expr: ((key < 15) and (key < 10))'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -410,7 +410,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '                filter predicates:'
 '                  0 '
 '                  1 '
-'                  2 {(key < 20.0)}'
+'                  2 {(key < 20)}'
 '                handleSkewJoin: false'
 '                keys:'
 '                  0 [Column[key]]'
@@ -435,7 +435,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '              filter predicates:'
 '                0 '
 '                1 '
-'                2 {(key < 20.0)}'
+'                2 {(key < 20)}'
 '              handleSkewJoin: false'
 '              keys:'
 '                0 [Column[key]]'
@@ -567,7 +567,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 10.0) and (key < 15.0))'
+'                  expr: ((key < 10) and (key < 15))'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -588,7 +588,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 15.0) and (key < 10.0))'
+'                  expr: ((key < 15) and (key < 10))'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -633,7 +633,7 @@ SORT BY k1,v1,k2,v2,k3,v3 
 '          filter predicates:'
 '            0 '
 '            1 '
-'            2 {(VALUE._col0 < 20.0)}'
+'            2 {(VALUE._col0 < 20)}'
 '          handleSkewJoin: false'
 '          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
 '          Select Operator'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out Tue Jul 30 22:22:35 2013
@@ -39,9 +39,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '                1 {key} {value}'
 '                2 {key} {value}'
 '              filter predicates:'
-'                0 {(key < 10.0)}'
+'                0 {(key < 10)}'
 '                1 '
-'                2 {(key < 10.0)}'
+'                2 {(key < 10)}'
 '              handleSkewJoin: false'
 '              keys:'
 '                0 [Column[key]]'
@@ -53,7 +53,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key > 10.0)'
+'                  expr: (key > 10)'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -61,9 +61,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '                  1 {key} {value}'
 '                  2 {key} {value}'
 '                filter predicates:'
-'                  0 {(key < 10.0)}'
+'                  0 {(key < 10)}'
 '                  1 '
-'                  2 {(key < 10.0)}'
+'                  2 {(key < 10)}'
 '                handleSkewJoin: false'
 '                keys:'
 '                  0 [Column[key]]'
@@ -86,9 +86,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '                1 {key} {value}'
 '                2 {key} {value}'
 '              filter predicates:'
-'                0 {(key < 10.0)}'
+'                0 {(key < 10)}'
 '                1 '
-'                2 {(key < 10.0)}'
+'                2 {(key < 10)}'
 '              handleSkewJoin: false'
 '              keys:'
 '                0 [Column[key]]'
@@ -187,7 +187,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key > 10.0)'
+'                  expr: (key > 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -230,9 +230,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            1 {VALUE._col0} {VALUE._col1}'
 '            2 {VALUE._col0} {VALUE._col1}'
 '          filter predicates:'
-'            0 {(VALUE._col0 < 10.0)}'
+'            0 {(VALUE._col0 < 10)}'
 '            1 '
-'            2 {(VALUE._col0 < 10.0)}'
+'            2 {(VALUE._col0 < 10)}'
 '          handleSkewJoin: false'
 '          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
 '          Select Operator'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out Tue Jul 30 22:22:35 2013
@@ -35,7 +35,7 @@ SELECT  *  FROM src src1 JOIN src src2 W
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -55,7 +55,7 @@ SELECT  *  FROM src src1 JOIN src src2 W
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Map Join Operator'
 '                condition map:'
@@ -135,7 +135,7 @@ SELECT  *  FROM src src1 JOIN src src2 W
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              HashTable Sink Operator'
 '                condition expressions:'
@@ -155,7 +155,7 @@ SELECT  *  FROM src src1 JOIN src src2 W
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Map Join Operator'
 '                condition map:'
@@ -197,7 +197,7 @@ SELECT  *  FROM src src1 JOIN src src2 W
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                sort order: '
@@ -212,7 +212,7 @@ SELECT  *  FROM src src1 JOIN src src2 W
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                sort order: '

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out Tue Jul 30 22:22:35 2013
@@ -38,7 +38,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -108,7 +108,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -160,7 +160,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -253,7 +253,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -281,7 +281,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -335,7 +335,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'
@@ -359,7 +359,7 @@ ON src_12.key = src3.k AND src3.k < 200;
 '            alias: src'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 200.0)'
+'                  expr: (key < 200)'
 '                  type: boolean'
 '              Select Operator'
 '                expressions:'

Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out Tue Jul 30 22:22:35 2013
@@ -46,7 +46,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key > 10.0)'
+'                  expr: (key > 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -89,9 +89,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            1 {VALUE._col0} {VALUE._col1}'
 '            2 {VALUE._col0} {VALUE._col1}'
 '          filter predicates:'
-'            0 {(VALUE._col0 < 10.0)}'
+'            0 {(VALUE._col0 < 10)}'
 '            1 '
-'            2 {(VALUE._col0 < 10.0)}'
+'            2 {(VALUE._col0 < 10)}'
 '          handleSkewJoin: false'
 '          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
 '          Select Operator'
@@ -206,7 +206,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            alias: src2'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key > 10.0)'
+'                  expr: (key > 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -227,7 +227,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            alias: src3'
 '            Filter Operator'
 '              predicate:'
-'                  expr: ((key < 10.0) and (key > 10.0))'
+'                  expr: ((key < 10) and (key > 10))'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -253,7 +253,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
 '            1 {VALUE._col0} {VALUE._col1}'
 '            2 {VALUE._col0} {VALUE._col1}'
 '          filter predicates:'
-'            0 {(VALUE._col0 < 10.0)}'
+'            0 {(VALUE._col0 < 10)}'
 '            1 '
 '            2 '
 '          handleSkewJoin: false'
@@ -353,7 +353,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN 
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -391,7 +391,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN 
 '            alias: src3'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -418,7 +418,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN 
 '            2 {VALUE._col0} {VALUE._col1}'
 '          filter predicates:'
 '            0 '
-'            1 {(VALUE._col0 > 10.0)}'
+'            1 {(VALUE._col0 > 10)}'
 '            2 '
 '          handleSkewJoin: false'
 '          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
@@ -517,7 +517,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN 
 '            alias: src1'
 '            Filter Operator'
 '              predicate:'
-'                  expr: (key < 10.0)'
+'                  expr: (key < 10)'
 '                  type: boolean'
 '              Reduce Output Operator'
 '                key expressions:'
@@ -578,8 +578,8 @@ SELECT * FROM src src1 RIGHT OUTER JOIN 
 '            2 {VALUE._col0} {VALUE._col1}'
 '          filter predicates:'
 '            0 '
-'            1 {(VALUE._col0 > 10.0)}'
-'            2 {(VALUE._col0 < 10.0)}'
+'            1 {(VALUE._col0 > 10)}'
+'            2 {(VALUE._col0 < 10)}'
 '          handleSkewJoin: false'
 '          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
 '          Select Operator'