You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2013/07/31 00:22:46 UTC
svn commit: r1508669 [7/39] - in /hive/branches/vectorization: ./
common/src/java/org/apache/hadoop/hive/conf/ conf/
contrib/src/test/results/clientpositive/ data/files/ eclipse-templates/
hcatalog/build-support/ant/ hcatalog/core/src/main/java/org/apa...
Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java Tue Jul 30 22:22:35 2013
@@ -76,6 +76,7 @@ public class GenericUDAFVariance extends
case DECIMAL:
return new GenericUDAFVarianceEvaluator();
case BOOLEAN:
+ case DATE:
default:
throw new UDFArgumentTypeException(0,
"Only numeric or string type arguments are accepted but "
Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVarianceSample.java Tue Jul 30 22:22:35 2013
@@ -59,6 +59,7 @@ public class GenericUDAFVarianceSample e
case DECIMAL:
return new GenericUDAFVarianceSampleEvaluator();
case BOOLEAN:
+ case DATE:
default:
throw new UDFArgumentTypeException(0,
"Only numeric or string type arguments are accepted but "
Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFnGrams.java Tue Jul 30 22:22:35 2013
@@ -26,22 +26,15 @@ import org.apache.hadoop.hive.ql.exec.De
import org.apache.hadoop.hive.ql.exec.UDFArgumentTypeException;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.ql.parse.SemanticException;
-import org.apache.hadoop.hive.serde2.io.DoubleWritable;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.StandardMapObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.StandardListObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.StructField;
-import org.apache.hadoop.hive.serde2.objectinspector.primitive.DoubleObjectInspector;
-import org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableDoubleObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils;
-import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo;
+import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-import org.apache.hadoop.util.StringUtils;
import org.apache.hadoop.io.Text;
/**
@@ -69,7 +62,7 @@ public class GenericUDAFnGrams implement
throw new UDFArgumentTypeException(parameters.length-1,
"Please specify either three or four arguments.");
}
-
+
// Validate the first parameter, which is the expression to compute over. This should be an
// array of strings type, or an array of arrays of strings.
PrimitiveTypeInfo pti;
@@ -99,7 +92,7 @@ public class GenericUDAFnGrams implement
}
if(pti.getPrimitiveCategory() != PrimitiveObjectInspector.PrimitiveCategory.STRING) {
throw new UDFArgumentTypeException(0,
- "Only array<string> or array<array<string>> is allowed, but "
+ "Only array<string> or array<array<string>> is allowed, but "
+ parameters[0].getTypeName() + " was passed as parameter 1.");
}
@@ -107,7 +100,7 @@ public class GenericUDAFnGrams implement
if(parameters[1].getCategory() != ObjectInspector.Category.PRIMITIVE) {
throw new UDFArgumentTypeException(1, "Only integers are accepted but "
+ parameters[1].getTypeName() + " was passed as parameter 2.");
- }
+ }
switch(((PrimitiveTypeInfo) parameters[1]).getPrimitiveCategory()) {
case BYTE:
case SHORT:
@@ -125,7 +118,7 @@ public class GenericUDAFnGrams implement
if(parameters[2].getCategory() != ObjectInspector.Category.PRIMITIVE) {
throw new UDFArgumentTypeException(2, "Only integers are accepted but "
+ parameters[2].getTypeName() + " was passed as parameter 3.");
- }
+ }
switch(((PrimitiveTypeInfo) parameters[2]).getPrimitiveCategory()) {
case BYTE:
case SHORT:
@@ -144,7 +137,7 @@ public class GenericUDAFnGrams implement
if(parameters[3].getCategory() != ObjectInspector.Category.PRIMITIVE) {
throw new UDFArgumentTypeException(3, "Only integers are accepted but "
+ parameters[3].getTypeName() + " was passed as parameter 4.");
- }
+ }
switch(((PrimitiveTypeInfo) parameters[3]).getPrimitiveCategory()) {
case BYTE:
case SHORT:
@@ -174,7 +167,7 @@ public class GenericUDAFnGrams implement
private PrimitiveObjectInspector kOI;
private PrimitiveObjectInspector pOI;
- // For PARTIAL2 and FINAL: ObjectInspectors for partial aggregations
+ // For PARTIAL2 and FINAL: ObjectInspectors for partial aggregations
private StandardListObjectInspector loi;
@Override
@@ -209,8 +202,8 @@ public class GenericUDAFnGrams implement
// Init output object inspectors.
//
// The return type for a partial aggregation is still a list of strings.
- //
- // The return type for FINAL and COMPLETE is a full aggregation result, which is
+ //
+ // The return type for FINAL and COMPLETE is a full aggregation result, which is
// an array of structures containing the n-gram and its estimated frequency.
if (m == Mode.PARTIAL1 || m == Mode.PARTIAL2) {
return ObjectInspectorFactory.getStandardListObjectInspector(
@@ -224,7 +217,7 @@ public class GenericUDAFnGrams implement
foi.add(PrimitiveObjectInspectorFactory.writableDoubleObjectInspector);
ArrayList<String> fname = new ArrayList<String>();
fname.add("ngram");
- fname.add("estfrequency");
+ fname.add("estfrequency");
return ObjectInspectorFactory.getStandardListObjectInspector(
ObjectInspectorFactory.getStandardStructObjectInspector(fname, foi) );
}
@@ -232,14 +225,14 @@ public class GenericUDAFnGrams implement
@Override
public void merge(AggregationBuffer agg, Object partial) throws HiveException {
- if(partial == null) {
+ if(partial == null) {
return;
}
NGramAggBuf myagg = (NGramAggBuf) agg;
List<Text> partialNGrams = (List<Text>) loi.getList(partial);
int n = Integer.parseInt(partialNGrams.get(partialNGrams.size()-1).toString());
if(myagg.n > 0 && myagg.n != n) {
- throw new HiveException(getClass().getSimpleName() + ": mismatch in value for 'n'"
+ throw new HiveException(getClass().getSimpleName() + ": mismatch in value for 'n'"
+ ", which usually is caused by a non-constant expression. Found '"+n+"' and '"
+ myagg.n + "'.");
}
@@ -273,7 +266,7 @@ public class GenericUDAFnGrams implement
return;
}
NGramAggBuf myagg = (NGramAggBuf) agg;
-
+
// Parse out 'n' and 'k' if we haven't already done so, and while we're at it,
// also parse out the precision factor 'pf' if the user has supplied one.
if(!myagg.nge.isInitialized()) {
@@ -317,7 +310,7 @@ public class GenericUDAFnGrams implement
// parse out n-grams, update frequency counts
processNgrams(myagg, words);
- }
+ }
} else {
// we're dealing with an array of strings
ArrayList<String> words = new ArrayList<String>();
@@ -337,7 +330,7 @@ public class GenericUDAFnGrams implement
return myagg.nge.getNGrams();
}
- // Aggregation buffer methods.
+ // Aggregation buffer methods.
static class NGramAggBuf extends AbstractAggregationBuffer {
NGramEstimator nge;
int n;
Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java Tue Jul 30 22:22:35 2013
@@ -136,10 +136,18 @@ public abstract class GenericUDFBaseComp
if (oiTypeInfo0 != oiTypeInfo1) {
compareType = CompareType.NEED_CONVERT;
- // If either argument is a string, we convert to a double because a number
- // in string form should always be convertible into a double
- if (oiTypeInfo0.equals(TypeInfoFactory.stringTypeInfo)
+ if ((oiTypeInfo0.equals(TypeInfoFactory.stringTypeInfo)
+ && oiTypeInfo1.equals(TypeInfoFactory.dateTypeInfo))
+ || (oiTypeInfo0.equals(TypeInfoFactory.dateTypeInfo)
+ && oiTypeInfo1.equals(TypeInfoFactory.stringTypeInfo))) {
+ // Date should be comparable with string
+ compareOI = TypeInfoUtils.getStandardWritableObjectInspectorFromTypeInfo(
+ TypeInfoFactory.stringTypeInfo);
+
+ } else if (oiTypeInfo0.equals(TypeInfoFactory.stringTypeInfo)
|| oiTypeInfo1.equals(TypeInfoFactory.stringTypeInfo)) {
+ // If either argument is a string, we convert to a double because a number
+ // in string form should always be convertible into a double
compareOI = TypeInfoUtils.getStandardWritableObjectInspectorFromTypeInfo(
TypeInfoFactory.doubleTypeInfo);
} else {
Modified: hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java (original)
+++ hive/branches/vectorization/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToUnixTimeStamp.java Tue Jul 30 22:22:35 2013
@@ -28,6 +28,7 @@ import org.apache.hadoop.hive.ql.exec.UD
import org.apache.hadoop.hive.ql.exec.UDFArgumentLengthException;
import org.apache.hadoop.hive.ql.metadata.HiveException;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
+import org.apache.hadoop.hive.serde2.objectinspector.primitive.DateObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.StringObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.TimestampObjectInspector;
@@ -42,6 +43,7 @@ import org.apache.hadoop.io.LongWritable
public class GenericUDFToUnixTimeStamp extends GenericUDF {
private StringObjectInspector intputTextOI;
+ private DateObjectInspector inputDateOI;
private TimestampObjectInspector inputTimestampOI;
private StringObjectInspector patternOI;
@@ -69,6 +71,8 @@ public class GenericUDFToUnixTimeStamp e
}
patternOI = (StringObjectInspector) arguments[1];
}
+ } else if (arguments[0] instanceof DateObjectInspector) {
+ inputDateOI = (DateObjectInspector) arguments[0];
} else if (arguments[0] instanceof TimestampObjectInspector) {
inputTimestampOI = (TimestampObjectInspector) arguments[0];
} else {
@@ -106,7 +110,11 @@ public class GenericUDFToUnixTimeStamp e
} catch (ParseException e) {
return null;
}
- }
+ } else if (inputDateOI != null) {
+ retValue.set(inputDateOI.getPrimitiveWritableObject(arguments[0].get())
+ .getTimeInSeconds());
+ return retValue;
+ }
Timestamp timestamp = inputTimestampOI.getPrimitiveJavaObject(arguments[0].get());
retValue.set(timestamp.getTime() / 1000);
return retValue;
Modified: hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto (original)
+++ hive/branches/vectorization/ql/src/protobuf/org/apache/hadoop/hive/ql/io/orc/orc_proto.proto Tue Jul 30 22:22:35 2013
@@ -27,6 +27,12 @@ message DecimalStatistics {
optional string sum = 3;
}
+message DateStatistics {
+ // min,max values saved as days since epoch
+ optional sint32 minimum = 1;
+ optional sint32 maximum = 2;
+}
+
message ColumnStatistics {
optional uint64 numberOfValues = 1;
optional IntegerStatistics intStatistics = 2;
@@ -34,6 +40,7 @@ message ColumnStatistics {
optional StringStatistics stringStatistics = 4;
optional BucketStatistics bucketStatistics = 5;
optional DecimalStatistics decimalStatistics = 6;
+ optional DateStatistics dateStatistics = 7;
}
message RowIndexEntry {
@@ -93,6 +100,7 @@ message Type {
STRUCT = 12;
UNION = 13;
DECIMAL = 14;
+ DATE = 15;
}
required Kind kind = 1;
repeated uint32 subtypes = 2 [packed=true];
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java Tue Jul 30 22:22:35 2013
@@ -886,7 +886,7 @@ public class QTestUtil {
FileOutputStream ofs = new FileOutputStream(outf);
for (Task<? extends Serializable> plan : tasks) {
- Utilities.serializeTasks(plan, ofs);
+ Utilities.serializeObject(plan, ofs);
}
String[] patterns = new String[] {
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestExecDriver.java Tue Jul 30 22:22:35 2013
@@ -52,6 +52,7 @@ import org.apache.hadoop.hive.ql.plan.Ma
import org.apache.hadoop.hive.ql.plan.PartitionDesc;
import org.apache.hadoop.hive.ql.plan.PlanUtils;
import org.apache.hadoop.hive.ql.plan.ReduceSinkDesc;
+import org.apache.hadoop.hive.ql.plan.ReduceWork;
import org.apache.hadoop.hive.ql.plan.ScriptDesc;
import org.apache.hadoop.hive.ql.plan.SelectDesc;
import org.apache.hadoop.hive.serde.serdeConstants;
@@ -141,7 +142,7 @@ public class TestExecDriver extends Test
}
public static void addMapWork(MapredWork mr, Table tbl, String alias, Operator<?> work) {
- mr.addMapWork(tbl.getDataLocation().toString(), alias, work, new PartitionDesc(
+ mr.getMapWork().addMapWork(tbl.getDataLocation().toString(), alias, work, new PartitionDesc(
Utilities.getTableDesc(tbl), null));
}
@@ -194,7 +195,6 @@ public class TestExecDriver extends Test
@SuppressWarnings("unchecked")
private void populateMapPlan1(Table src) {
- mr.setNumReduceTasks(Integer.valueOf(0));
Operator<FileSinkDesc> op2 = OperatorFactory.get(new FileSinkDesc(tmpdir
+ "mapplan1.out", Utilities.defaultTd, true));
@@ -206,7 +206,6 @@ public class TestExecDriver extends Test
@SuppressWarnings("unchecked")
private void populateMapPlan2(Table src) {
- mr.setNumReduceTasks(Integer.valueOf(0));
Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
+ "mapplan2.out", Utilities.defaultTd, false));
@@ -225,7 +224,6 @@ public class TestExecDriver extends Test
@SuppressWarnings("unchecked")
private void populateMapRedPlan1(Table src) throws SemanticException {
- mr.setNumReduceTasks(Integer.valueOf(1));
ArrayList<String> outputColumns = new ArrayList<String>();
for (int i = 0; i < 2; i++) {
@@ -238,8 +236,11 @@ public class TestExecDriver extends Test
-1, 1, -1));
addMapWork(mr, src, "a", op1);
- mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
- mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ ReduceWork rWork = new ReduceWork();
+ rWork.setNumReduceTasks(Integer.valueOf(1));
+ rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+ rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ mr.setReduceWork(rWork);
// reduce side work
Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -248,12 +249,11 @@ public class TestExecDriver extends Test
Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
- mr.setReducer(op2);
+ rWork.setReducer(op2);
}
@SuppressWarnings("unchecked")
private void populateMapRedPlan2(Table src) throws SemanticException {
- mr.setNumReduceTasks(Integer.valueOf(1));
ArrayList<String> outputColumns = new ArrayList<String>();
for (int i = 0; i < 2; i++) {
outputColumns.add("_col" + i);
@@ -266,8 +266,11 @@ public class TestExecDriver extends Test
outputColumns, false, -1, 1, -1));
addMapWork(mr, src, "a", op1);
- mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
- mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ ReduceWork rWork = new ReduceWork();
+ rWork.setNumReduceTasks(Integer.valueOf(1));
+ rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+ rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ mr.setReduceWork(rWork);
// reduce side work
Operator<FileSinkDesc> op4 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -278,7 +281,7 @@ public class TestExecDriver extends Test
Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
- mr.setReducer(op2);
+ rWork.setReducer(op2);
}
/**
@@ -286,8 +289,6 @@ public class TestExecDriver extends Test
*/
@SuppressWarnings("unchecked")
private void populateMapRedPlan3(Table src, Table src2) throws SemanticException {
- mr.setNumReduceTasks(Integer.valueOf(5));
- mr.setNeedsTagging(true);
List<String> outputColumns = new ArrayList<String>();
for (int i = 0; i < 2; i++) {
outputColumns.add("_col" + i);
@@ -299,8 +300,6 @@ public class TestExecDriver extends Test
Byte.valueOf((byte) 0), 1, -1));
addMapWork(mr, src, "a", op1);
- mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
- mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
Operator<ReduceSinkDesc> op2 = OperatorFactory.get(PlanUtils
.getReduceSinkDesc(Utilities.makeList(getStringColumn("key")),
@@ -308,7 +307,14 @@ public class TestExecDriver extends Test
Byte.valueOf((byte) 1), Integer.MAX_VALUE, -1));
addMapWork(mr, src2, "b", op2);
- mr.getTagToValueDesc().add(op2.getConf().getValueSerializeInfo());
+ ReduceWork rWork = new ReduceWork();
+ rWork.setNumReduceTasks(Integer.valueOf(5));
+ rWork.setNeedsTagging(true);
+ rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+ rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+
+ mr.setReduceWork(rWork);
+ rWork.getTagToValueDesc().add(op2.getConf().getValueSerializeInfo());
// reduce side work
Operator<FileSinkDesc> op4 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -320,12 +326,11 @@ public class TestExecDriver extends Test
Utilities.ReduceField.VALUE.toString(), "", false), "0", false)),
Utilities.makeList(outputColumns.get(0))), op4);
- mr.setReducer(op5);
+ rWork.setReducer(op5);
}
@SuppressWarnings("unchecked")
private void populateMapRedPlan4(Table src) throws SemanticException {
- mr.setNumReduceTasks(Integer.valueOf(1));
// map-side work
ArrayList<String> outputColumns = new ArrayList<String>();
@@ -348,8 +353,11 @@ public class TestExecDriver extends Test
outputColumns), op0);
addMapWork(mr, src, "a", op4);
- mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
- mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ ReduceWork rWork = new ReduceWork();
+ rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+ rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ rWork.setNumReduceTasks(Integer.valueOf(1));
+ mr.setReduceWork(rWork);
// reduce side work
Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -358,7 +366,7 @@ public class TestExecDriver extends Test
Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
- mr.setReducer(op2);
+ rWork.setReducer(op2);
}
public static ExprNodeColumnDesc getStringColumn(String columnName) {
@@ -368,7 +376,6 @@ public class TestExecDriver extends Test
@SuppressWarnings("unchecked")
private void populateMapRedPlan5(Table src) throws SemanticException {
- mr.setNumReduceTasks(Integer.valueOf(1));
// map-side work
ArrayList<String> outputColumns = new ArrayList<String>();
@@ -385,8 +392,11 @@ public class TestExecDriver extends Test
outputColumns), op0);
addMapWork(mr, src, "a", op4);
- mr.setKeyDesc(op0.getConf().getKeySerializeInfo());
- mr.getTagToValueDesc().add(op0.getConf().getValueSerializeInfo());
+ ReduceWork rWork = new ReduceWork();
+ mr.setReduceWork(rWork);
+ rWork.setNumReduceTasks(Integer.valueOf(1));
+ rWork.setKeyDesc(op0.getConf().getKeySerializeInfo());
+ rWork.getTagToValueDesc().add(op0.getConf().getValueSerializeInfo());
// reduce side work
Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -395,12 +405,11 @@ public class TestExecDriver extends Test
Operator<ExtractDesc> op2 = OperatorFactory.get(new ExtractDesc(
getStringColumn(Utilities.ReduceField.VALUE.toString())), op3);
- mr.setReducer(op2);
+ rWork.setReducer(op2);
}
@SuppressWarnings("unchecked")
private void populateMapRedPlan6(Table src) throws SemanticException {
- mr.setNumReduceTasks(Integer.valueOf(1));
// map-side work
ArrayList<String> outputColumns = new ArrayList<String>();
@@ -424,8 +433,11 @@ public class TestExecDriver extends Test
outputColumns), op0);
addMapWork(mr, src, "a", op4);
- mr.setKeyDesc(op1.getConf().getKeySerializeInfo());
- mr.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
+ ReduceWork rWork = new ReduceWork();
+ mr.setReduceWork(rWork);
+ rWork.setNumReduceTasks(Integer.valueOf(1));
+ rWork.setKeyDesc(op1.getConf().getKeySerializeInfo());
+ rWork.getTagToValueDesc().add(op1.getConf().getValueSerializeInfo());
// reduce side work
Operator<FileSinkDesc> op3 = OperatorFactory.get(new FileSinkDesc(tmpdir
@@ -436,7 +448,7 @@ public class TestExecDriver extends Test
Operator<ExtractDesc> op5 = OperatorFactory.get(new ExtractDesc(
getStringColumn(Utilities.ReduceField.VALUE.toString())), op2);
- mr.setReducer(op5);
+ rWork.setReducer(op5);
}
private void executePlan() throws Exception {
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java Tue Jul 30 22:22:35 2013
@@ -18,20 +18,20 @@
package org.apache.hadoop.hive.ql.exec;
-import java.util.List;
-import java.util.LinkedList;
import java.lang.reflect.Method;
+import java.util.LinkedList;
+import java.util.List;
import junit.framework.TestCase;
-import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-import org.apache.hadoop.hive.serde.serdeConstants;
-import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.serde2.io.DoubleWritable;
import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable;
-import org.apache.hadoop.io.IntWritable;
-import org.apache.hadoop.io.BytesWritable;
import org.apache.hadoop.hive.serde2.io.TimestampWritable;
+import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
+import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
+import org.apache.hadoop.io.BytesWritable;
+import org.apache.hadoop.io.IntWritable;
public class TestFunctionRegistry extends TestCase {
@@ -41,10 +41,11 @@ public class TestFunctionRegistry extend
public void one(IntWritable x, HiveDecimalWritable y) {}
public void one(IntWritable x, DoubleWritable y) {}
public void one(IntWritable x, IntWritable y) {}
+ public void mismatch(DateWritable x, HiveDecimalWritable y) {}
public void mismatch(TimestampWritable x, HiveDecimalWritable y) {}
public void mismatch(BytesWritable x, DoubleWritable y) {}
}
-
+
@Override
protected void setUp() {
}
@@ -58,17 +59,18 @@ public class TestFunctionRegistry extend
implicit(TypeInfoFactory.floatTypeInfo, TypeInfoFactory.decimalTypeInfo, true);
implicit(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo, true);
implicit(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo, true);
+ implicit(TypeInfoFactory.dateTypeInfo, TypeInfoFactory.decimalTypeInfo, false);
implicit(TypeInfoFactory.timestampTypeInfo, TypeInfoFactory.decimalTypeInfo, false);
}
- private void verify(Class udf, String name, TypeInfo ta, TypeInfo tb,
+ private void verify(Class udf, String name, TypeInfo ta, TypeInfo tb,
Class a, Class b, boolean throwException) {
List<TypeInfo> args = new LinkedList<TypeInfo>();
args.add(ta);
args.add(tb);
Method result = null;
-
+
try {
result = FunctionRegistry.getMethodInternal(udf, name, false, args);
} catch (UDFArgumentException e) {
@@ -116,13 +118,13 @@ public class TestFunctionRegistry extend
}
public void testCommonClass() {
- common(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo,
+ common(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo,
TypeInfoFactory.decimalTypeInfo);
- common(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo,
+ common(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo,
TypeInfoFactory.stringTypeInfo);
- common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo,
+ common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo,
TypeInfoFactory.decimalTypeInfo);
- common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo,
+ common(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo,
TypeInfoFactory.stringTypeInfo);
}
@@ -131,13 +133,13 @@ public class TestFunctionRegistry extend
}
public void testCommonClassComparison() {
- comparison(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo,
+ comparison(TypeInfoFactory.intTypeInfo, TypeInfoFactory.decimalTypeInfo,
TypeInfoFactory.decimalTypeInfo);
- comparison(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo,
+ comparison(TypeInfoFactory.stringTypeInfo, TypeInfoFactory.decimalTypeInfo,
TypeInfoFactory.decimalTypeInfo);
- comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo,
+ comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.decimalTypeInfo,
TypeInfoFactory.decimalTypeInfo);
- comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo,
+ comparison(TypeInfoFactory.doubleTypeInfo, TypeInfoFactory.stringTypeInfo,
TypeInfoFactory.doubleTypeInfo);
}
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperators.java Tue Jul 30 22:22:35 2013
@@ -345,13 +345,13 @@ public class TestOperators extends TestC
// initialize mapredWork
MapredWork mrwork = new MapredWork();
- mrwork.setPathToAliases(pathToAliases);
- mrwork.setPathToPartitionInfo(pathToPartitionInfo);
- mrwork.setAliasToWork(aliasToWork);
+ mrwork.getMapWork().setPathToAliases(pathToAliases);
+ mrwork.getMapWork().setPathToPartitionInfo(pathToPartitionInfo);
+ mrwork.getMapWork().setAliasToWork(aliasToWork);
// get map operator and initialize it
MapOperator mo = new MapOperator();
- mo.initializeAsRoot(hconf, mrwork);
+ mo.initializeAsRoot(hconf, mrwork.getMapWork());
Text tw = new Text();
InspectableObject io1 = new InspectableObject();
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestPlan.java Tue Jul 30 22:22:35 2013
@@ -75,13 +75,13 @@ public class TestPlan extends TestCase {
ao.put("a", op);
MapredWork mrwork = new MapredWork();
- mrwork.setPathToAliases(pa);
- mrwork.setPathToPartitionInfo(pt);
- mrwork.setAliasToWork(ao);
+ mrwork.getMapWork().setPathToAliases(pa);
+ mrwork.getMapWork().setPathToPartitionInfo(pt);
+ mrwork.getMapWork().setAliasToWork(ao);
// serialize the configuration once ..
ByteArrayOutputStream baos = new ByteArrayOutputStream();
- Utilities.serializeMapRedWork(mrwork, baos);
+ Utilities.serializeObject(mrwork, baos);
baos.close();
String v1 = baos.toString();
@@ -91,7 +91,7 @@ public class TestPlan extends TestCase {
Utilities.setMapRedWork(job, mrwork, System.getProperty("java.io.tmpdir") + File.separator +
System.getProperty("user.name") + File.separator + "hive");
MapredWork mrwork2 = Utilities.getMapRedWork(job);
- Utilities.clearMapRedWork(job);
+ Utilities.clearWork(job);
// over here we should have some checks of the deserialized object against
// the orginal object
@@ -99,7 +99,7 @@ public class TestPlan extends TestCase {
// serialize again
baos.reset();
- Utilities.serializeMapRedWork(mrwork2, baos);
+ Utilities.serializeObject(mrwork2, baos);
baos.close();
// verify that the two are equal
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/exec/TestUtilities.java Tue Jul 30 22:22:35 2013
@@ -18,11 +18,18 @@
package org.apache.hadoop.hive.ql.exec;
+import java.sql.Date;
+import java.sql.Timestamp;
+
import static org.apache.hadoop.hive.ql.exec.Utilities.getFileExtension;
import junit.framework.TestCase;
+import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat;
+import org.apache.hadoop.hive.ql.plan.ExprNodeConstantDesc;
+import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
+import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
import org.apache.hadoop.mapred.JobConf;
public class TestUtilities extends TestCase {
@@ -53,4 +60,14 @@ public class TestUtilities extends TestC
assertEquals("Custom extension for uncompressed text format", extension,
getFileExtension(jc, true, new HiveIgnoreKeyTextOutputFormat()));
}
+
+ public void testSerializeTimestamp() {
+ Timestamp ts = new Timestamp(1374554702000L);
+ ts.setNanos(123456);
+ ExprNodeConstantDesc constant = new ExprNodeConstantDesc(
+ TypeInfoFactory.timestampTypeInfo, ts);
+ String serialized = Utilities.serializeExpression(constant);
+ ExprNodeDesc deserialized = Utilities.deserializeExpression(serialized, new Configuration());
+ assertEquals(constant.getExprString(), deserialized.getExprString());
+ }
}
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/hooks/VerifyHiveSortedInputFormatUsedHook.java Tue Jul 30 22:22:35 2013
@@ -38,7 +38,7 @@ public class VerifyHiveSortedInputFormat
for (Task<? extends Serializable> rootTask : rootTasks) {
if (rootTask.getWork() instanceof MapredWork) {
Assert.assertTrue("The root map reduce task's input was not marked as sorted.",
- ((MapredWork)rootTask.getWork()).isInputFormatSorted());
+ ((MapredWork)rootTask.getWork()).getMapWork().isInputFormatSorted());
}
}
}
Modified: hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java (original)
+++ hive/branches/vectorization/ql/src/test/org/apache/hadoop/hive/ql/io/TestSymlinkTextInputFormat.java Tue Jul 30 22:22:35 2013
@@ -46,7 +46,6 @@ import org.apache.hadoop.mapred.InputSpl
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.RecordReader;
import org.apache.hadoop.mapred.Reporter;
-import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.util.ReflectionUtils;
/**
@@ -167,8 +166,8 @@ public class TestSymlinkTextInputFormat
QueryPlan plan = drv.getPlan();
MapRedTask selectTask = (MapRedTask)plan.getRootTasks().get(0);
-
- ExecDriver.addInputPaths(newJob, selectTask.getWork(), emptyScratchDir.toString(), ctx);
+
+ ExecDriver.addInputPaths(newJob, selectTask.getWork().getMapWork(), emptyScratchDir.toString(), ctx);
Utilities.setMapRedWork(newJob, selectTask.getWork(), ctx.getMRTmpFileURI());
CombineHiveInputFormat combineInputFormat = ReflectionUtils.newInstance(
Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter1.q Tue Jul 30 22:22:35 2013
@@ -1,2 +1,2 @@
CREATE TABLE alter_test (d STRING);
-ALTER TABLE alter_test CHANGE d d DATE;
+ALTER TABLE alter_test CHANGE d d DATETIME;
Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_alter2.q Tue Jul 30 22:22:35 2013
@@ -1,2 +1,2 @@
CREATE TABLE alter_test (d STRING);
-ALTER TABLE alter_test ADD COLUMNS (ds DATE);
+ALTER TABLE alter_test ADD COLUMNS (ds DATETIME);
Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/invalid_t_transform.q Tue Jul 30 22:22:35 2013
@@ -1 +1 @@
-SELECT TRANSFORM(*) USING 'cat' AS (key DATE) FROM src;
+SELECT TRANSFORM(*) USING 'cat' AS (key DATETIME) FROM src;
Modified: hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientnegative/serde_regex.q Tue Jul 30 22:22:35 2013
@@ -1,5 +1,5 @@
USE default;
--- This should fail because Regex SerDe doesn't support TIMESTAMP, STRUCT
+-- This should fail because Regex SerDe doesn't support STRUCT
CREATE TABLE serde_regex(
host STRING,
identity STRING,
Modified: hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q (original)
+++ hive/branches/vectorization/ql/src/test/queries/clientpositive/avro_nullable_fields.q Tue Jul 30 22:22:35 2013
@@ -12,8 +12,8 @@ CREATE TABLE test_serializer(string1 STR
struct1 STRUCT<sint:INT,sboolean:BOOLEAN,sstring:STRING>,
enum1 STRING,
nullableint INT,
- bytes1 ARRAY<TINYINT>,
- fixed1 ARRAY<TINYINT>)
+ bytes1 BINARY,
+ fixed1 BINARY)
ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' COLLECTION ITEMS TERMINATED BY ':' MAP KEYS TERMINATED BY '#' LINES TERMINATED BY '\n'
STORED AS TEXTFILE;
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join0.q.out Tue Jul 30 22:22:35 2013
@@ -45,7 +45,7 @@ SORT BY k1, v1, k2, v2
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -72,7 +72,7 @@ SORT BY k1, v1, k2, v2
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -205,7 +205,7 @@ SORT BY k1, v1, k2, v2
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -232,7 +232,7 @@ SORT BY k1, v1, k2, v2
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -281,7 +281,7 @@ SORT BY k1, v1, k2, v2
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -303,7 +303,7 @@ SORT BY k1, v1, k2, v2
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Select Operator'
' expressions:'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join12.q.out Tue Jul 30 22:22:35 2013
@@ -51,7 +51,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -76,7 +76,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 80.0) and (key < 100.0))'
+' expr: ((key < 80) and (key < 100))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -103,7 +103,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -191,7 +191,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -214,7 +214,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 80.0) and (key < 100.0))'
+' expr: ((key < 80) and (key < 100))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -241,7 +241,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -302,7 +302,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -325,7 +325,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -354,7 +354,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 80.0) and (key < 100.0))'
+' expr: ((key < 80) and (key < 100))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -406,7 +406,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -430,7 +430,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 100.0) and (key < 80.0))'
+' expr: ((key < 100) and (key < 80))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -456,7 +456,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: ((key < 80.0) and (key < 100.0))'
+' expr: ((key < 80) and (key < 100))'
' type: boolean'
' Select Operator'
' expressions:'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join13.q.out Tue Jul 30 22:22:35 2013
@@ -50,7 +50,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 100.0)'
+' expr: (key < 100)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -77,7 +77,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 100.0)'
+' expr: (key < 100)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -120,7 +120,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -230,7 +230,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -294,7 +294,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -351,7 +351,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 100.0)'
+' expr: (key < 100)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -376,7 +376,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 100.0)'
+' expr: (key < 100)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -414,7 +414,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 100.0)'
+' expr: (key < 100)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -438,7 +438,7 @@ ON src1.c1 + src2.c3 = src3.c5 AND src3.
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 100.0)'
+' expr: (key < 100)'
' type: boolean'
' Select Operator'
' expressions:'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join16.q.out Tue Jul 30 22:22:35 2013
@@ -42,7 +42,7 @@ where tab.value < 200;
' alias: tab'
' Filter Operator'
' predicate:'
-' expr: ((key > 20.0) and (value < 200.0))'
+' expr: ((key > 20) and (value < 200))'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -62,7 +62,7 @@ where tab.value < 200;
' alias: a'
' Filter Operator'
' predicate:'
-' expr: (((key > 10.0) and (key > 20.0)) and (value < 200.0))'
+' expr: (((key > 10) and (key > 20)) and (value < 200))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -146,7 +146,7 @@ where tab.value < 200;
' alias: a'
' Filter Operator'
' predicate:'
-' expr: (((key > 10.0) and (key > 20.0)) and (value < 200.0))'
+' expr: (((key > 10) and (key > 20)) and (value < 200))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -173,7 +173,7 @@ where tab.value < 200;
' alias: tab'
' Filter Operator'
' predicate:'
-' expr: ((key > 20.0) and (value < 200.0))'
+' expr: ((key > 20) and (value < 200))'
' type: boolean'
' Map Join Operator'
' condition map:'
@@ -217,7 +217,7 @@ where tab.value < 200;
' alias: a'
' Filter Operator'
' predicate:'
-' expr: (((key > 10.0) and (key > 20.0)) and (value < 200.0))'
+' expr: (((key > 10) and (key > 20)) and (value < 200))'
' type: boolean'
' Select Operator'
' expressions:'
@@ -247,7 +247,7 @@ where tab.value < 200;
' alias: tab'
' Filter Operator'
' predicate:'
-' expr: ((key > 20.0) and (value < 200.0))'
+' expr: ((key > 20) and (value < 200))'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join20.q.out Tue Jul 30 22:22:35 2013
@@ -42,7 +42,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -52,7 +52,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(key < 20.0)}'
+' 2 {(key < 20)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -64,7 +64,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -74,7 +74,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(key < 20.0)}'
+' 2 {(key < 20)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -99,7 +99,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(key < 20.0)}'
+' 2 {(key < 20)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -231,7 +231,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -252,7 +252,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -297,7 +297,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(VALUE._col0 < 20.0)}'
+' 2 {(VALUE._col0 < 20)}'
' handleSkewJoin: false'
' outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
' Select Operator'
@@ -378,7 +378,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: ((key < 10.0) and (key < 15.0))'
+' expr: ((key < 10) and (key < 15))'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -388,7 +388,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(key < 20.0)}'
+' 2 {(key < 20)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -400,7 +400,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: ((key < 15.0) and (key < 10.0))'
+' expr: ((key < 15) and (key < 10))'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -410,7 +410,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(key < 20.0)}'
+' 2 {(key < 20)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -435,7 +435,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(key < 20.0)}'
+' 2 {(key < 20)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -567,7 +567,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: ((key < 10.0) and (key < 15.0))'
+' expr: ((key < 10) and (key < 15))'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -588,7 +588,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: ((key < 15.0) and (key < 10.0))'
+' expr: ((key < 15) and (key < 10))'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -633,7 +633,7 @@ SORT BY k1,v1,k2,v2,k3,v3
' filter predicates:'
' 0 '
' 1 '
-' 2 {(VALUE._col0 < 20.0)}'
+' 2 {(VALUE._col0 < 20)}'
' handleSkewJoin: false'
' outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
' Select Operator'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join21.q.out Tue Jul 30 22:22:35 2013
@@ -39,9 +39,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' 1 {key} {value}'
' 2 {key} {value}'
' filter predicates:'
-' 0 {(key < 10.0)}'
+' 0 {(key < 10)}'
' 1 '
-' 2 {(key < 10.0)}'
+' 2 {(key < 10)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -53,7 +53,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key > 10.0)'
+' expr: (key > 10)'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -61,9 +61,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' 1 {key} {value}'
' 2 {key} {value}'
' filter predicates:'
-' 0 {(key < 10.0)}'
+' 0 {(key < 10)}'
' 1 '
-' 2 {(key < 10.0)}'
+' 2 {(key < 10)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -86,9 +86,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' 1 {key} {value}'
' 2 {key} {value}'
' filter predicates:'
-' 0 {(key < 10.0)}'
+' 0 {(key < 10)}'
' 1 '
-' 2 {(key < 10.0)}'
+' 2 {(key < 10)}'
' handleSkewJoin: false'
' keys:'
' 0 [Column[key]]'
@@ -187,7 +187,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key > 10.0)'
+' expr: (key > 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -230,9 +230,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' 1 {VALUE._col0} {VALUE._col1}'
' 2 {VALUE._col0} {VALUE._col1}'
' filter predicates:'
-' 0 {(VALUE._col0 < 10.0)}'
+' 0 {(VALUE._col0 < 10)}'
' 1 '
-' 2 {(VALUE._col0 < 10.0)}'
+' 2 {(VALUE._col0 < 10)}'
' handleSkewJoin: false'
' outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
' Select Operator'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join23.q.out Tue Jul 30 22:22:35 2013
@@ -35,7 +35,7 @@ SELECT * FROM src src1 JOIN src src2 W
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -55,7 +55,7 @@ SELECT * FROM src src1 JOIN src src2 W
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Map Join Operator'
' condition map:'
@@ -135,7 +135,7 @@ SELECT * FROM src src1 JOIN src src2 W
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' HashTable Sink Operator'
' condition expressions:'
@@ -155,7 +155,7 @@ SELECT * FROM src src1 JOIN src src2 W
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Map Join Operator'
' condition map:'
@@ -197,7 +197,7 @@ SELECT * FROM src src1 JOIN src src2 W
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' sort order: '
@@ -212,7 +212,7 @@ SELECT * FROM src src1 JOIN src src2 W
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' sort order: '
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join27.q.out Tue Jul 30 22:22:35 2013
@@ -38,7 +38,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -108,7 +108,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -160,7 +160,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -253,7 +253,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -281,7 +281,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -335,7 +335,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
@@ -359,7 +359,7 @@ ON src_12.key = src3.k AND src3.k < 200;
' alias: src'
' Filter Operator'
' predicate:'
-' expr: (key < 200.0)'
+' expr: (key < 200)'
' type: boolean'
' Select Operator'
' expressions:'
Modified: hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out (original)
+++ hive/branches/vectorization/ql/src/test/results/beelinepositive/auto_join28.q.out Tue Jul 30 22:22:35 2013
@@ -46,7 +46,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key > 10.0)'
+' expr: (key > 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -89,9 +89,9 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' 1 {VALUE._col0} {VALUE._col1}'
' 2 {VALUE._col0} {VALUE._col1}'
' filter predicates:'
-' 0 {(VALUE._col0 < 10.0)}'
+' 0 {(VALUE._col0 < 10)}'
' 1 '
-' 2 {(VALUE._col0 < 10.0)}'
+' 2 {(VALUE._col0 < 10)}'
' handleSkewJoin: false'
' outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
' Select Operator'
@@ -206,7 +206,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' alias: src2'
' Filter Operator'
' predicate:'
-' expr: (key > 10.0)'
+' expr: (key > 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -227,7 +227,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' alias: src3'
' Filter Operator'
' predicate:'
-' expr: ((key < 10.0) and (key > 10.0))'
+' expr: ((key < 10) and (key > 10))'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -253,7 +253,7 @@ SELECT * FROM src src1 LEFT OUTER JOIN s
' 1 {VALUE._col0} {VALUE._col1}'
' 2 {VALUE._col0} {VALUE._col1}'
' filter predicates:'
-' 0 {(VALUE._col0 < 10.0)}'
+' 0 {(VALUE._col0 < 10)}'
' 1 '
' 2 '
' handleSkewJoin: false'
@@ -353,7 +353,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -391,7 +391,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN
' alias: src3'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -418,7 +418,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN
' 2 {VALUE._col0} {VALUE._col1}'
' filter predicates:'
' 0 '
-' 1 {(VALUE._col0 > 10.0)}'
+' 1 {(VALUE._col0 > 10)}'
' 2 '
' handleSkewJoin: false'
' outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
@@ -517,7 +517,7 @@ SELECT * FROM src src1 RIGHT OUTER JOIN
' alias: src1'
' Filter Operator'
' predicate:'
-' expr: (key < 10.0)'
+' expr: (key < 10)'
' type: boolean'
' Reduce Output Operator'
' key expressions:'
@@ -578,8 +578,8 @@ SELECT * FROM src src1 RIGHT OUTER JOIN
' 2 {VALUE._col0} {VALUE._col1}'
' filter predicates:'
' 0 '
-' 1 {(VALUE._col0 > 10.0)}'
-' 2 {(VALUE._col0 < 10.0)}'
+' 1 {(VALUE._col0 > 10)}'
+' 2 {(VALUE._col0 < 10)}'
' handleSkewJoin: false'
' outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9'
' Select Operator'