You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by mm...@apache.org on 2018/06/25 10:38:40 UTC
[32/33] hive git commit: Revert "HIVE-12192 : Hive should carry out
timestamp computations in UTC (Jesus Camacho Rodriguez via Ashutosh Chauhan)"
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hbase-handler/src/test/results/positive/hbase_timestamp.q.out
----------------------------------------------------------------------
diff --git a/hbase-handler/src/test/results/positive/hbase_timestamp.q.out b/hbase-handler/src/test/results/positive/hbase_timestamp.q.out
index 6e7d7e6..fabbfba 100644
--- a/hbase-handler/src/test/results/positive/hbase_timestamp.q.out
+++ b/hbase-handler/src/test/results/positive/hbase_timestamp.q.out
@@ -97,26 +97,26 @@ POSTHOOK: query: SELECT key, value, cast(`time` as timestamp) FROM hbase_table
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hbase_table
#### A masked pattern was here ####
-0 val_0 2012-02-23 01:15:54
-119 val_119 2012-02-23 01:15:54
-136 val_136 2012-02-23 01:15:54
-153 val_153 2012-02-23 01:15:54
-17 val_17 2012-02-23 01:15:54
-170 val_170 2012-02-23 01:15:54
-187 val_187 2012-02-23 01:15:54
-221 val_221 2012-02-23 01:15:54
-238 val_238 2012-02-23 01:15:54
-255 val_255 2012-02-23 01:15:54
-272 val_272 2012-02-23 01:15:54
-289 val_289 2012-02-23 01:15:54
-306 val_306 2012-02-23 01:15:54
-323 val_323 2012-02-23 01:15:54
-34 val_34 2012-02-23 01:15:54
-374 val_374 2012-02-23 01:15:54
-459 val_459 2012-02-23 01:15:54
-493 val_493 2012-02-23 01:15:54
-51 val_51 2012-02-23 01:15:54
-85 val_85 2012-02-23 01:15:54
+0 val_0 2012-02-22 17:15:54
+119 val_119 2012-02-22 17:15:54
+136 val_136 2012-02-22 17:15:54
+153 val_153 2012-02-22 17:15:54
+17 val_17 2012-02-22 17:15:54
+170 val_170 2012-02-22 17:15:54
+187 val_187 2012-02-22 17:15:54
+221 val_221 2012-02-22 17:15:54
+238 val_238 2012-02-22 17:15:54
+255 val_255 2012-02-22 17:15:54
+272 val_272 2012-02-22 17:15:54
+289 val_289 2012-02-22 17:15:54
+306 val_306 2012-02-22 17:15:54
+323 val_323 2012-02-22 17:15:54
+34 val_34 2012-02-22 17:15:54
+374 val_374 2012-02-22 17:15:54
+459 val_459 2012-02-22 17:15:54
+493 val_493 2012-02-22 17:15:54
+51 val_51 2012-02-22 17:15:54
+85 val_85 2012-02-22 17:15:54
PREHOOK: query: DROP TABLE hbase_table
PREHOOK: type: DROPTABLE
PREHOOK: Input: default@hbase_table
@@ -202,8 +202,8 @@ POSTHOOK: query: SELECT key, value, cast(`time` as timestamp) FROM hbase_table W
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hbase_table
#### A masked pattern was here ####
-165 val_165 1973-03-03 09:46:40
-396 val_396 1973-03-03 09:46:40
+165 val_165 1973-03-03 01:46:40
+396 val_396 1973-03-03 01:46:40
PREHOOK: query: explain
SELECT key, value, cast(`time` as timestamp) FROM hbase_table WHERE key > 100 AND key < 400 AND `time` > 100000000000
PREHOOK: type: QUERY
@@ -251,10 +251,10 @@ POSTHOOK: query: SELECT key, value, cast(`time` as timestamp) FROM hbase_table W
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hbase_table
#### A masked pattern was here ####
-111 val_111 1976-05-03 19:33:20
-222 val_222 1976-05-03 19:33:20
-296 val_296 1976-05-03 19:33:20
-333 val_333 1976-05-03 19:33:20
+111 val_111 1976-05-03 12:33:20
+222 val_222 1976-05-03 12:33:20
+296 val_296 1976-05-03 12:33:20
+333 val_333 1976-05-03 12:33:20
PREHOOK: query: explain
SELECT key, value, cast(`time` as timestamp) FROM hbase_table WHERE key > 100 AND key < 400 AND `time` <= 100000000000
PREHOOK: type: QUERY
@@ -302,8 +302,8 @@ POSTHOOK: query: SELECT key, value, cast(`time` as timestamp) FROM hbase_table W
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hbase_table
#### A masked pattern was here ####
-165 val_165 1973-03-03 09:46:40
-396 val_396 1973-03-03 09:46:40
+165 val_165 1973-03-03 01:46:40
+396 val_396 1973-03-03 01:46:40
PREHOOK: query: explain
SELECT key, value, cast(`time` as timestamp) FROM hbase_table WHERE key > 100 AND key < 400 AND `time` >= 200000000000
PREHOOK: type: QUERY
@@ -351,10 +351,10 @@ POSTHOOK: query: SELECT key, value, cast(`time` as timestamp) FROM hbase_table W
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hbase_table
#### A masked pattern was here ####
-111 val_111 1976-05-03 19:33:20
-222 val_222 1976-05-03 19:33:20
-296 val_296 1976-05-03 19:33:20
-333 val_333 1976-05-03 19:33:20
+111 val_111 1976-05-03 12:33:20
+222 val_222 1976-05-03 12:33:20
+296 val_296 1976-05-03 12:33:20
+333 val_333 1976-05-03 12:33:20
PREHOOK: query: DROP TABLE hbase_table
PREHOOK: type: DROPTABLE
PREHOOK: Input: default@hbase_table
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/DataType.java
----------------------------------------------------------------------
diff --git a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/DataType.java b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/DataType.java
index d33d343..6dcee40 100644
--- a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/DataType.java
+++ b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/DataType.java
@@ -19,12 +19,12 @@
package org.apache.hive.hcatalog.data;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java
----------------------------------------------------------------------
diff --git a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java
index 405f1b0..57c48f8 100644
--- a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java
+++ b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java
@@ -19,16 +19,16 @@
package org.apache.hive.hcatalog.data;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.List;
import java.util.Map;
import org.apache.hadoop.hive.common.classification.InterfaceAudience;
import org.apache.hadoop.hive.common.classification.InterfaceStability;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hive.hcatalog.common.HCatException;
import org.apache.hive.hcatalog.data.schema.HCatSchema;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/JsonSerDe.java
----------------------------------------------------------------------
diff --git a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/JsonSerDe.java b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/JsonSerDe.java
index af80c02..114c205 100644
--- a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/JsonSerDe.java
+++ b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/JsonSerDe.java
@@ -21,6 +21,8 @@ package org.apache.hive.hcatalog.data;
import java.io.ByteArrayInputStream;
import java.io.IOException;
import java.nio.charset.CharacterCodingException;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collections;
@@ -32,11 +34,9 @@ import java.util.regex.Matcher;
import java.util.regex.Pattern;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.serde.serdeConstants;
import org.apache.hadoop.hive.serde2.AbstractSerDe;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/ReaderWriter.java
----------------------------------------------------------------------
diff --git a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/ReaderWriter.java b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/ReaderWriter.java
index 2641add..cb1c459 100644
--- a/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/ReaderWriter.java
+++ b/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/ReaderWriter.java
@@ -22,6 +22,7 @@ package org.apache.hive.hcatalog.data;
import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
+import java.sql.Date;
import java.util.ArrayList;
import java.util.LinkedHashMap;
import java.util.Iterator;
@@ -29,16 +30,14 @@ import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.serde2.io.HiveCharWritable;
import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable;
import org.apache.hadoop.hive.serde2.io.HiveVarcharWritable;
-import org.apache.hadoop.hive.serde2.io.TimestampWritableV2;
+import org.apache.hadoop.hive.serde2.io.TimestampWritable;
import org.apache.hadoop.io.VIntWritable;
import org.apache.hadoop.io.VLongWritable;
@@ -122,11 +121,11 @@ public abstract class ReaderWriter {
hdw.readFields(in);
return hdw.getHiveDecimal();
case DataType.DATE:
- DateWritableV2 dw = new DateWritableV2();
+ DateWritable dw = new DateWritable();
dw.readFields(in);
return dw.get();
case DataType.TIMESTAMP:
- TimestampWritableV2 tw = new TimestampWritableV2();
+ TimestampWritable tw = new TimestampWritable();
tw.readFields(in);
return tw.getTimestamp();
default:
@@ -215,10 +214,10 @@ public abstract class ReaderWriter {
new HiveDecimalWritable((HiveDecimal)val).write(out);
return;
case DataType.DATE:
- new DateWritableV2((Date)val).write(out);
+ new DateWritable((Date)val).write(out);
return;
case DataType.TIMESTAMP:
- new TimestampWritableV2((Timestamp)val).write(out);
+ new TimestampWritable((java.sql.Timestamp)val).write(out);
return;
default:
throw new IOException("Unexpected data type " + type +
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestDefaultHCatRecord.java
----------------------------------------------------------------------
diff --git a/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestDefaultHCatRecord.java b/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestDefaultHCatRecord.java
index d57d171..7a643f1 100644
--- a/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestDefaultHCatRecord.java
+++ b/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestDefaultHCatRecord.java
@@ -29,17 +29,17 @@ import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;
import java.math.BigDecimal;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Calendar;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hive.hcatalog.common.HCatException;
import org.apache.hive.hcatalog.data.schema.HCatSchema;
import org.apache.hive.hcatalog.data.schema.HCatSchemaUtils;
@@ -276,7 +276,7 @@ public class TestDefaultHCatRecord extends TestCase {
rec_hcat13types.add(new HiveChar("hive_char", 10));
rec_hcat13types.add(new HiveVarchar("hive_varchar", 20));
rec_hcat13types.add(Date.valueOf("2014-01-06"));
- rec_hcat13types.add(Timestamp.ofEpochMilli(System.currentTimeMillis()));
+ rec_hcat13types.add(new Timestamp(System.currentTimeMillis()));
return new DefaultHCatRecord(rec_hcat13types);
}
private static HCatRecord getHCat13TypesComplexRecord() {
@@ -290,7 +290,7 @@ public class TestDefaultHCatRecord extends TestCase {
List<Object> list = new ArrayList<Object>();
list.add(Date.valueOf("2014-01-05"));
list.add(new HashMap<HiveDecimal, String>(m));
- m2.put(Timestamp.ofEpochMilli(System.currentTimeMillis()), list);
+ m2.put(new Timestamp(System.currentTimeMillis()), list);
rec_hcat13ComplexTypes.add(m2);
return new DefaultHCatRecord(rec_hcat13ComplexTypes);
}
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestJsonSerDe.java
----------------------------------------------------------------------
diff --git a/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestJsonSerDe.java b/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestJsonSerDe.java
index 6770d44..8aeb4f4 100644
--- a/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestJsonSerDe.java
+++ b/hcatalog/core/src/test/java/org/apache/hive/hcatalog/data/TestJsonSerDe.java
@@ -20,6 +20,8 @@ package org.apache.hive.hcatalog.data;
import java.io.UnsupportedEncodingException;
import java.math.BigDecimal;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashMap;
@@ -30,11 +32,9 @@ import java.util.Properties;
import junit.framework.TestCase;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.serde.serdeConstants;
import org.apache.hadoop.hive.serde2.SerDeUtils;
@@ -98,7 +98,7 @@ public class TestJsonSerDe extends TestCase {
rlist.add(new HiveChar("hive\nchar", 10));
rlist.add(new HiveVarchar("hive\nvarchar", 20));
rlist.add(Date.valueOf("2014-01-07"));
- rlist.add(Timestamp.ofEpochMilli(System.currentTimeMillis()));
+ rlist.add(new Timestamp(System.currentTimeMillis()));
rlist.add("hive\nbinary".getBytes("UTF-8"));
List<Object> nlist = new ArrayList<Object>(13);
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java
----------------------------------------------------------------------
diff --git a/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java b/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java
index 994c505..ec620d2 100644
--- a/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java
+++ b/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatBaseStorer.java
@@ -21,6 +21,8 @@ package org.apache.hive.hcatalog.pig;
import java.io.IOException;
import java.math.BigDecimal;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Collections;
import java.util.HashMap;
@@ -33,11 +35,9 @@ import java.util.Properties;
import org.apache.commons.lang.StringUtils;
import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hadoop.hive.serde2.typeinfo.CharTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
@@ -419,7 +419,7 @@ abstract class HCatBaseStorer extends StoreFunc implements StoreMetadata {
return new HiveVarchar(varcharVal, vti.getLength());
case TIMESTAMP:
DateTime dt = (DateTime)pigObj;
- return Timestamp.ofEpochMilli(dt.getMillis());//toEpochMilli() returns UTC time regardless of TZ
+ return new Timestamp(dt.getMillis());//getMillis() returns UTC time regardless of TZ
case DATE:
/**
* We ignore any TZ setting on Pig value since java.sql.Date doesn't have it (in any
@@ -437,7 +437,7 @@ abstract class HCatBaseStorer extends StoreFunc implements StoreMetadata {
for local timezone. Date.valueOf() also uses local timezone (as does Date(int,int,int).
Also see PigHCatUtil#extractPigObject() for corresponding read op. This way a DATETIME from Pig,
when stored into Hive and read back comes back with the same value.*/
- return Date.of(dateTime.getYear(), dateTime.getMonthOfYear(), dateTime.getDayOfMonth());
+ return new Date(dateTime.getYear() - 1900, dateTime.getMonthOfYear() - 1, dateTime.getDayOfMonth());
default:
throw new BackendException("Unexpected HCat type " + type + " for value " + pigObj
+ " of class " + pigObj.getClass().getName(), PigHCatUtil.PIG_EXCEPTION_CODE);
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java
----------------------------------------------------------------------
diff --git a/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java b/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java
index c3bde2d..f5c3c75 100644
--- a/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java
+++ b/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java
@@ -54,7 +54,6 @@ import org.apache.pig.ResourceSchema;
import org.apache.pig.ResourceStatistics;
import org.apache.pig.impl.util.UDFContext;
import org.joda.time.DateTime;
-import org.joda.time.DateTimeZone;
import org.joda.time.format.DateTimeFormat;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java
----------------------------------------------------------------------
diff --git a/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java b/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java
index afe6e92..163fe6a 100644
--- a/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java
+++ b/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java
@@ -20,6 +20,8 @@ package org.apache.hive.hcatalog.pig;
import java.io.IOException;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashMap;
@@ -29,11 +31,9 @@ import java.util.Map.Entry;
import java.util.Properties;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hive.common.type.Date;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.common.type.HiveVarchar;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.metastore.IMetaStoreClient;
import org.apache.hadoop.hive.metastore.Warehouse;
@@ -62,7 +62,6 @@ import org.apache.pig.impl.logicalLayer.schema.Schema;
import org.apache.pig.impl.util.UDFContext;
import org.apache.pig.impl.util.Utils;
import org.joda.time.DateTime;
-import org.joda.time.DateTimeZone;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
@@ -430,7 +429,7 @@ class PigHCatUtil {
* e.g. d = new java.sql.Date(System.currentMillis()).toString() so if you do this just after
* midnight in Palo Alto, you'll get yesterday's date printed out.*/
Date d = (Date)o;
- result = new DateTime(d.getYear(), d.getMonth(), d.getDay(), 0, 0, DateTimeZone.UTC);
+ result = new DateTime(d.getYear() + 1900, d.getMonth() + 1, d.getDate(), 0, 0);//uses local TZ
break;
case TIMESTAMP:
/*DATA TRUNCATION!!!
@@ -438,7 +437,7 @@ class PigHCatUtil {
object in local TZ; This is arbitrary, since Hive value doesn't have any TZ notion, but
we need to set something for TZ.
Timestamp is consistently in GMT (unless you call toString() on it) so we use millis*/
- result = new DateTime(((Timestamp)o).toEpochMilli(), DateTimeZone.UTC);
+ result = new DateTime(((Timestamp)o).getTime());//uses local TZ
break;
default:
result = o;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatLoaderTest.java
----------------------------------------------------------------------
diff --git a/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatLoaderTest.java b/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatLoaderTest.java
index 58981f8..0d72102 100644
--- a/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatLoaderTest.java
+++ b/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatLoaderTest.java
@@ -29,6 +29,8 @@ import java.io.FileWriter;
import java.io.IOException;
import java.io.PrintWriter;
import java.io.RandomAccessFile;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Collection;
import java.util.HashMap;
@@ -40,8 +42,6 @@ import java.util.Properties;
import org.apache.commons.io.FileUtils;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hive.common.type.Date;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hadoop.hive.ql.IDriver;
import org.apache.hadoop.hive.ql.processors.CommandProcessorResponse;
import org.apache.hadoop.hive.serde2.ColumnProjectionUtils;
@@ -652,7 +652,7 @@ public abstract class AbstractHCatLoaderTest extends HCatBaseTest {
* All the values are within range of target data type (column)
*/
private static final Object[][] primitiveRows = new Object[][] {
- {Boolean.TRUE,Byte.MAX_VALUE,Short.MAX_VALUE, Integer.MAX_VALUE,Long.MAX_VALUE,Float.MAX_VALUE,Double.MAX_VALUE,555.22,"Kyiv","char(10)xx","varchar(20)","blah".getBytes(), Date.valueOf("2014-01-13"), Timestamp.valueOf("2014-01-13 19:26:25.0123")},
+ {Boolean.TRUE,Byte.MAX_VALUE,Short.MAX_VALUE, Integer.MAX_VALUE,Long.MAX_VALUE,Float.MAX_VALUE,Double.MAX_VALUE,555.22,"Kyiv","char(10)xx","varchar(20)","blah".getBytes(),Date.valueOf("2014-01-13"),Timestamp.valueOf("2014-01-13 19:26:25.0123")},
{Boolean.FALSE,Byte.MIN_VALUE,Short.MIN_VALUE, Integer.MIN_VALUE,Long.MIN_VALUE,Float.MIN_VALUE,Double.MIN_VALUE,-555.22,"Saint Petersburg","char(xx)00","varchar(yy)","doh".getBytes(),Date.valueOf("2014-01-14"), Timestamp.valueOf("2014-01-14 19:26:25.0123")}
};
/**
@@ -701,22 +701,14 @@ public abstract class AbstractHCatLoaderTest extends HCatBaseTest {
assertTrue("rowNum=" + numTuplesRead + " colNum=" + colPos
+ " Reference data is null; actual "
+ t.get(colPos), t.get(colPos) == null);
- } else if (referenceData instanceof Date) {
+ } else if (referenceData instanceof java.util.Date) {
// Note that here we ignore nanos part of Hive Timestamp since nanos are dropped when
// reading Hive from Pig by design.
assertTrue("rowNum=" + numTuplesRead + " colNum=" + colPos
- + " Reference data=" + ((Date)referenceData).toEpochMilli()
- + " actual=" + ((DateTime)t.get(colPos)).getMillis()
- + "; types=(" + referenceData.getClass() + "," + t.get(colPos).getClass() + ")",
- ((Date)referenceData).toEpochMilli() == ((DateTime)t.get(colPos)).getMillis());
- } else if (referenceData instanceof Timestamp) {
- // Note that here we ignore nanos part of Hive Timestamp since nanos are dropped when
- // reading Hive from Pig by design.
- assertTrue("rowNum=" + numTuplesRead + " colNum=" + colPos
- + " Reference data=" + ((Timestamp)referenceData).toEpochMilli()
+ + " Reference data=" + ((java.util.Date)referenceData).getTime()
+ " actual=" + ((DateTime)t.get(colPos)).getMillis()
+ "; types=(" + referenceData.getClass() + "," + t.get(colPos).getClass() + ")",
- ((Timestamp)referenceData).toEpochMilli()== ((DateTime)t.get(colPos)).getMillis());
+ ((java.util.Date)referenceData).getTime()== ((DateTime)t.get(colPos)).getMillis());
} else {
// Doing String comps here as value objects in Hive in Pig are different so equals()
// doesn't work.
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatStorerTest.java
----------------------------------------------------------------------
diff --git a/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatStorerTest.java b/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatStorerTest.java
index 19c30b0..a5cf3a5 100644
--- a/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatStorerTest.java
+++ b/hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/AbstractHCatStorerTest.java
@@ -213,17 +213,23 @@ public abstract class AbstractHCatStorerTest extends HCatBaseTest {
*/
@Test
public void testWriteTimestamp() throws Exception {
- DateTime d = new DateTime(1991, 10, 11, 14, 23, 30, 10, DateTimeZone.UTC);// uses default TZ
+ DateTime d = new DateTime(1991, 10, 11, 14, 23, 30, 10);// uses default TZ
pigValueRangeTest("junitTypeTest1", "timestamp", "datetime", null, d.toString(),
- d.toDateTime(DateTimeZone.UTC).toString());
+ d.toDateTime(DateTimeZone.getDefault()).toString());
d = d.plusHours(2);
pigValueRangeTest("junitTypeTest2", "timestamp", "datetime",
HCatBaseStorer.OOR_VALUE_OPT_VALUES.Null, d.toString(),
- d.toDateTime(DateTimeZone.UTC).toString());
+ d.toDateTime(DateTimeZone.getDefault()).toString());
+ d = d.toDateTime(DateTimeZone.UTC);
+ pigValueRangeTest("junitTypeTest3", "timestamp", "datetime", null, d.toString(),
+ d.toDateTime(DateTimeZone.getDefault()).toString());
- d = new DateTime(1991, 10, 11, 23, 24, 25, 26, DateTimeZone.UTC);
+ d = new DateTime(1991, 10, 11, 23, 24, 25, 26);
pigValueRangeTest("junitTypeTest1", "timestamp", "datetime", null, d.toString(),
- d.toDateTime(DateTimeZone.UTC).toString());
+ d.toDateTime(DateTimeZone.getDefault()).toString());
+ d = d.toDateTime(DateTimeZone.UTC);
+ pigValueRangeTest("junitTypeTest3", "timestamp", "datetime", null, d.toString(),
+ d.toDateTime(DateTimeZone.getDefault()).toString());
}
// End: tests that check values from Pig that are out of range for target column
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/ColumnVectorGenUtil.java
----------------------------------------------------------------------
diff --git a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/ColumnVectorGenUtil.java b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/ColumnVectorGenUtil.java
index 2ed5e5e..d80b6d4 100644
--- a/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/ColumnVectorGenUtil.java
+++ b/itests/hive-jmh/src/main/java/org/apache/hive/benchmark/vectorization/ColumnVectorGenUtil.java
@@ -16,10 +16,11 @@
package org.apache.hive.benchmark.vectorization;
+import java.sql.Timestamp;
import java.util.Random;
import org.apache.hadoop.hive.common.type.HiveDecimal;
-import org.apache.hadoop.hive.common.type.Timestamp;
+import org.apache.hadoop.hive.common.type.RandomTypeUtil;
import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.ColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.DecimalColumnVector;
@@ -27,14 +28,12 @@ import org.apache.hadoop.hive.ql.exec.vector.DoubleColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.TimestampColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
-import org.apache.hadoop.hive.serde2.RandomTypeUtil;
import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-
public class ColumnVectorGenUtil {
private static final long LONG_VECTOR_NULL_VALUE = 1;
@@ -145,7 +144,7 @@ public class ColumnVectorGenUtil {
final boolean repeating, final int size, final Random rand) {
Timestamp[] timestamps = new Timestamp[size];
for (int i = 0; i < size; i++) {
- timestamps[i] = Timestamp.ofEpochMilli(rand.nextInt());
+ timestamps[i] = new Timestamp(rand.nextInt());
}
return generateTimestampColumnVector(nulls, repeating, size, rand, timestamps);
}
@@ -170,10 +169,10 @@ public class ColumnVectorGenUtil {
tcv.isNull[i] = false;
if (!repeating) {
Timestamp randomTimestamp = RandomTypeUtil.getRandTimestamp(rand);
- tcv.set(i, randomTimestamp.toSqlTimestamp());
+ tcv.set(i, randomTimestamp);
timestampValues[i] = randomTimestamp;
} else {
- tcv.set(i, repeatingTimestamp.toSqlTimestamp());
+ tcv.set(i, repeatingTimestamp);
timestampValues[i] = repeatingTimestamp;
}
}
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/itests/hive-unit/src/test/java/org/apache/hive/jdbc/BaseJdbcWithMiniLlap.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/BaseJdbcWithMiniLlap.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/BaseJdbcWithMiniLlap.java
index 280119b..7a891ef 100644
--- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/BaseJdbcWithMiniLlap.java
+++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/BaseJdbcWithMiniLlap.java
@@ -31,10 +31,12 @@ import java.lang.reflect.Field;
import java.math.BigDecimal;
import java.net.URL;
import java.sql.Connection;
+import java.sql.Date;
import java.sql.DriverManager;
import java.sql.ResultSet;
import java.sql.SQLException;
import java.sql.Statement;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashMap;
@@ -74,8 +76,6 @@ import org.apache.hadoop.io.Text;
import org.apache.hive.jdbc.miniHS2.MiniHS2;
import org.apache.hive.jdbc.miniHS2.MiniHS2.MiniClusterType;
-import org.apache.hadoop.hive.common.type.Date;
-import org.apache.hadoop.hive.common.type.Timestamp;
import org.apache.hadoop.hive.llap.LlapBaseInputFormat;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector.Category;
import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector.PrimitiveCategory;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniLlapArrow.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniLlapArrow.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniLlapArrow.java
index e69c686..9dfece9 100644
--- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniLlapArrow.java
+++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniLlapArrow.java
@@ -21,8 +21,8 @@ package org.apache.hive.jdbc;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertArrayEquals;
import java.math.BigDecimal;
-import org.apache.hadoop.hive.common.type.Date;
-import org.apache.hadoop.hive.common.type.Timestamp;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.List;
import org.apache.hadoop.hive.llap.FieldDesc;
import org.apache.hadoop.hive.llap.Row;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/itests/src/test/resources/testconfiguration.properties
----------------------------------------------------------------------
diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties
index 2b36ceb..517b413 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -1686,7 +1686,6 @@ druid.query.files=druidmini_test1.q,\
druidmini_test_insert.q,\
druidmini_mv.q,\
druid_timestamptz.q,\
- druid_timestamptz2.q,\
druidmini_dynamic_partition.q,\
druidmini_expressions.q,\
druidmini_extractTime.q,\
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/llap-common/src/test/org/apache/hadoop/hive/llap/io/TestChunkedInputStream.java
----------------------------------------------------------------------
diff --git a/llap-common/src/test/org/apache/hadoop/hive/llap/io/TestChunkedInputStream.java b/llap-common/src/test/org/apache/hadoop/hive/llap/io/TestChunkedInputStream.java
index f328d6e..77559e1 100644
--- a/llap-common/src/test/org/apache/hadoop/hive/llap/io/TestChunkedInputStream.java
+++ b/llap-common/src/test/org/apache/hadoop/hive/llap/io/TestChunkedInputStream.java
@@ -18,6 +18,8 @@
package org.apache.hadoop.hive.llap.io;
+import java.io.ByteArrayInputStream;
+import java.io.ByteArrayOutputStream;
import java.io.FilterInputStream;
import java.io.FilterOutputStream;
import java.io.IOException;
@@ -29,7 +31,7 @@ import java.util.Arrays;
import java.util.List;
import java.util.Random;
-import org.apache.hadoop.hive.serde2.RandomTypeUtil;
+import org.apache.hadoop.hive.common.type.RandomTypeUtil;
import org.junit.Test;
import static org.junit.Assert.*;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/GenericColumnVectorProducer.java
----------------------------------------------------------------------
diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/GenericColumnVectorProducer.java b/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/GenericColumnVectorProducer.java
index af853e3..32f3bed 100644
--- a/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/GenericColumnVectorProducer.java
+++ b/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/GenericColumnVectorProducer.java
@@ -105,7 +105,7 @@ public class GenericColumnVectorProducer implements ColumnVectorProducer {
public static final class SerDeStripeMetadata implements ConsumerStripeMetadata {
// The writer is local to the process.
- private final String writerTimezone = "UTC";
+ private final String writerTimezone = TimeZone.getDefault().getID();
private List<ColumnEncoding> encodings;
private final int stripeIx;
private long rowCount = -1;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/OrcEncodedDataConsumer.java
----------------------------------------------------------------------
diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/OrcEncodedDataConsumer.java b/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/OrcEncodedDataConsumer.java
index 40248a3..0d7435c 100644
--- a/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/OrcEncodedDataConsumer.java
+++ b/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/OrcEncodedDataConsumer.java
@@ -223,8 +223,7 @@ public class OrcEncodedDataConsumer
TreeReaderFactory.Context context = new TreeReaderFactory.ReaderContext()
.setSchemaEvolution(evolution).skipCorrupt(skipCorrupt)
.writerTimeZone(stripeMetadata.getWriterTimezone())
- .fileFormat(fileMetadata == null ? null : fileMetadata.getFileVersion())
- .useUTCTimestamp(true);
+ .fileFormat(fileMetadata == null ? null : fileMetadata.getFileVersion());
this.batchSchemas = includes.getBatchReaderTypes(fileSchema);
StructTreeReader treeReader = EncodedTreeReaderFactory.createRootTreeReader(
batchSchemas, stripeMetadata.getEncodings(), batch, codec, context, useDecimal64ColumnVectors);
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DTIColumnCompareScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DTIColumnCompareScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/DTIColumnCompareScalar.txt
index f4e85bd..0d3ee2b 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DTIColumnCompareScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DTIColumnCompareScalar.txt
@@ -21,8 +21,8 @@ package org.apache.hadoop.hive.ql.exec.vector.expressions.gen;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
-import org.apache.hadoop.hive.common.type.Date;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
+import java.sql.Date;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
/**
* Generated from template DTIColumnCompareScalar.txt, which covers comparison
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DTIScalarCompareColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DTIScalarCompareColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/DTIScalarCompareColumn.txt
index b198e08..be5f641 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DTIScalarCompareColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DTIScalarCompareColumn.txt
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.exec.vector.expressions.gen;
-import org.apache.hadoop.hive.common.type.Date;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
+import java.sql.Date;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression;
import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthColumn.txt
index c3982ed..32dd6ed 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthColumn.txt
@@ -26,8 +26,8 @@ import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateColumnArithmeticIntervalYearMonthColumn.txt, which covers binary arithmetic
@@ -99,65 +99,65 @@ public class <ClassName> extends VectorExpression {
* conditional checks in the inner loop.
*/
if (inputColVector1.isRepeating && inputColVector2.isRepeating) {
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[0]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[0]));
scratchIntervalYearMonth2.set((int) vector2[0]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[0] = DateWritableV2.dateToDays(outputDate);
+ outputVector[0] = DateWritable.dateToDays(outputDate);
} else if (inputColVector1.isRepeating) {
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[0]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[0]));
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int i = 0; i != n; i++) {
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else if (inputColVector2.isRepeating) {
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int i = 0; i != n; i++) {
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else {
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int i = 0; i != n; i++) {
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
scratchDate1, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
}
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthScalar.txt
index 06c7368..94c0c5c 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticIntervalYearMonthScalar.txt
@@ -29,8 +29,8 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateColumnArithmeticIntervalYearMonthScalar.txt, which covers binary arithmetic
@@ -93,10 +93,10 @@ public class <ClassName> extends VectorExpression {
if (inputColVector1.isRepeating) {
if (inputColVector1.noNulls || !inputIsNull[0]) {
outputIsNull[0] = false;
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[0]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[0]));
dtm.<OperatorMethod>(
scratchDate1, value, outputDate);
- outputVector[0] = DateWritableV2.dateToDays(outputDate);
+ outputVector[0] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[0] = true;
outputColVector.noNulls = false;
@@ -115,18 +115,18 @@ public class <ClassName> extends VectorExpression {
for(int j = 0; j != n; j++) {
final int i = sel[j];
outputIsNull[i] = false;
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchDate1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int j = 0; j != n; j++) {
final int i = sel[j];
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchDate1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else {
@@ -138,10 +138,10 @@ public class <ClassName> extends VectorExpression {
outputColVector.noNulls = true;
}
for(int i = 0; i != n; i++) {
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchDate1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else /* there are NULLs in the inputColVector */ {
@@ -155,10 +155,10 @@ public class <ClassName> extends VectorExpression {
int i = sel[j];
if (!inputIsNull[i]) {
outputIsNull[i] = false;
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchDate1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
@@ -168,10 +168,10 @@ public class <ClassName> extends VectorExpression {
for(int i = 0; i != n; i++) {
if (!inputIsNull[i]) {
outputIsNull[i] = false;
- scratchDate1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchDate1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchDate1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampColumn.txt
index 53637a6..96c525d 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampColumn.txt
@@ -28,8 +28,8 @@ import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateColumnArithmeticTimestampColumn.txt, a class
@@ -97,12 +97,12 @@ public class <ClassName> extends VectorExpression {
* conditional checks in the inner loop.
*/
if (inputColVector1.isRepeating && inputColVector2.isRepeating) {
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[0]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[0]));
dtm.<OperatorMethod>(
scratchTimestamp1, inputColVector2.asScratch<CamelOperandType2>(0), outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(0);
} else if (inputColVector1.isRepeating) {
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[0]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[0]));
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
@@ -122,14 +122,14 @@ public class <ClassName> extends VectorExpression {
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
}
} else {
for(int i = 0; i != n; i++) {
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
@@ -139,14 +139,14 @@ public class <ClassName> extends VectorExpression {
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, inputColVector2.asScratch<CamelOperandType2>(i), outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
}
} else {
for(int i = 0; i != n; i++) {
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, inputColVector2.asScratch<CamelOperandType2>(i), outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampScalar.txt
index e9e9193..fb22992 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DateColumnArithmeticTimestampScalar.txt
@@ -29,8 +29,8 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateColumnArithmeticTimestampScalarBase.txt, a base class
@@ -91,7 +91,7 @@ public class <ClassName> extends VectorExpression {
if (inputColVector1.isRepeating) {
if (inputColVector1.noNulls || !inputIsNull[0]) {
outputIsNull[0] = false;
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[0]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[0]));
dtm.<OperatorMethod>(
scratchTimestamp1, value, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(0);
@@ -112,7 +112,7 @@ public class <ClassName> extends VectorExpression {
for(int j = 0; j != n; j++) {
final int i = sel[j];
outputIsNull[i] = false;
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
@@ -120,7 +120,7 @@ public class <ClassName> extends VectorExpression {
} else {
for(int j = 0; j != n; j++) {
final int i = sel[j];
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
@@ -135,7 +135,7 @@ public class <ClassName> extends VectorExpression {
outputColVector.noNulls = true;
}
for(int i = 0; i != n; i++) {
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
@@ -152,7 +152,7 @@ public class <ClassName> extends VectorExpression {
int i = sel[j];
if (!inputIsNull[i]) {
outputIsNull[i] = false;
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
@@ -165,7 +165,7 @@ public class <ClassName> extends VectorExpression {
for(int i = 0; i != n; i++) {
if (!inputIsNull[i]) {
outputIsNull[i] = false;
- scratchTimestamp1.setTime(DateWritableV2.daysToMillis((int) vector1[i]));
+ scratchTimestamp1.setTime(DateWritable.daysToMillis((int) vector1[i]));
dtm.<OperatorMethod>(
scratchTimestamp1, value, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
@@ -182,7 +182,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return getColumnParamString(0, colNum) + ", val " + TimestampUtils.timestampScalarTypeToString(value);
+ return getColumnParamString(0, colNum) + ", val " + value.toString();
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticIntervalYearMonthColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticIntervalYearMonthColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticIntervalYearMonthColumn.txt
index 7d2434a..0c8ec9c 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticIntervalYearMonthColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticIntervalYearMonthColumn.txt
@@ -37,8 +37,8 @@ import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateTimeScalarArithmeticIntervalYearMonthColumn.txt.
@@ -58,7 +58,7 @@ public class <ClassName> extends VectorExpression {
public <ClassName>(long value, int colNum, int outputColumnNum) {
super(outputColumnNum);
- this.value = new Date(DateWritableV2.daysToMillis((int) value));
+ this.value = new Date(DateWritable.daysToMillis((int) value));
this.colNum = colNum;
}
@@ -110,7 +110,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth2.set((int) vector2[0]);
dtm.<OperatorMethod>(
value, scratchIntervalYearMonth2, outputDate);
- outputVector[0] = DateWritableV2.dateToDays(outputDate);
+ outputVector[0] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[0] = true;
outputColVector.noNulls = false;
@@ -131,7 +131,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
value, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int j = 0; j != n; j++) {
@@ -139,7 +139,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
value, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else {
@@ -154,7 +154,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
value, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else /* there are NULLs in the inputColVector */ {
@@ -171,7 +171,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
value, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
@@ -184,7 +184,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth2.set((int) vector2[i]);
dtm.<OperatorMethod>(
value, scratchIntervalYearMonth2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticTimestampColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticTimestampColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticTimestampColumn.txt
index 678d827..ef8f2a3 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticTimestampColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/DateScalarArithmeticTimestampColumn.txt
@@ -36,8 +36,8 @@ import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateTimeScalarArithmeticTimestampColumnBase.txt.
@@ -57,7 +57,7 @@ public class <ClassName> extends VectorExpression {
super(outputColumnNum);
// Scalar input #1 is type date (days). For the math we convert it to a timestamp.
this.value = new Timestamp(0);
- this.value.setTime(DateWritableV2.daysToMillis((int) value));
+ this.value.setTime(DateWritable.daysToMillis((int) value));
this.colNum = colNum;
}
@@ -189,7 +189,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return "val " + org.apache.hadoop.hive.common.type.Date.ofEpochMilli(value.getTime()) + ", " + getColumnParamString(1, colNum);
+ return "val " + value + ", " + getColumnParamString(1, colNum);
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnBetweenDynamicValue.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnBetweenDynamicValue.txt b/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnBetweenDynamicValue.txt
index 1785abe..5242bbd 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnBetweenDynamicValue.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnBetweenDynamicValue.txt
@@ -22,7 +22,7 @@ import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.expressions.gen.Filter<TypeName>ColumnBetween;
import org.apache.hadoop.hive.ql.plan.DynamicValue;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorUtils;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampColumnCompareTimestampScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampColumnCompareTimestampScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampColumnCompareTimestampScalar.txt
index 6a05d77..3fa542b 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampColumnCompareTimestampScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampColumnCompareTimestampScalar.txt
@@ -141,7 +141,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return getColumnParamString(0, colNum) + ", val " + TimestampUtils.timestampScalarTypeToString(value);
+ return getColumnParamString(0, colNum) + ", val " + value.toString();
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampScalarCompareTimestampColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampScalarCompareTimestampColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampScalarCompareTimestampColumn.txt
index 36628a7..fcda65f 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampScalarCompareTimestampColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/FilterTimestampScalarCompareTimestampColumn.txt
@@ -144,7 +144,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return "val " + TimestampUtils.timestampScalarTypeToString(value) + ", " + getColumnParamString(1, colNum);
+ return "val " + value.toString() + ", " + getColumnParamString(1, colNum);
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateColumn.txt
index 4473bf0..393413f 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateColumn.txt
@@ -26,8 +26,8 @@ import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateColumnArithmeticIntervalYearMonthColumn.txt, which covers binary arithmetic
@@ -100,44 +100,44 @@ public class <ClassName> extends VectorExpression {
*/
if (inputColVector1.isRepeating && inputColVector2.isRepeating) {
scratchIntervalYearMonth1.set((int) vector1[0]);
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[0]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[0]));
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[0] = DateWritableV2.dateToDays(outputDate);
+ outputVector[0] = DateWritable.dateToDays(outputDate);
} else if (inputColVector1.isRepeating) {
scratchIntervalYearMonth1.set((int) vector1[0]);
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int i = 0; i != n; i++) {
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else if (inputColVector2.isRepeating) {
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[0]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[0]));
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int i = 0; i != n; i++) {
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else {
@@ -145,18 +145,18 @@ public class <ClassName> extends VectorExpression {
for(int j = 0; j != n; j++) {
int i = sel[j];
scratchIntervalYearMonth1.set((int) vector1[i]);
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int i = 0; i != n; i++) {
scratchIntervalYearMonth1.set((int) vector1[i]);
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
}
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateScalar.txt
index f0c1910..a65c4b9 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticDateScalar.txt
@@ -29,8 +29,8 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateColumnArithmeticIntervalYearMonthScalar.txt, which covers binary arithmetic
@@ -50,7 +50,7 @@ public class <ClassName> extends VectorExpression {
public <ClassName>(int colNum, long value, int outputColumnNum) {
super(outputColumnNum);
this.colNum = colNum;
- this.value = new Date(DateWritableV2.daysToMillis((int) value));
+ this.value = new Date(DateWritable.daysToMillis((int) value));
}
public <ClassName>() {
@@ -96,7 +96,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth1.set((int) vector1[0]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, value, outputDate);
- outputVector[0] = DateWritableV2.dateToDays(outputDate);
+ outputVector[0] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[0] = true;
outputColVector.noNulls = false;
@@ -118,7 +118,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int j = 0; j != n; j++) {
@@ -126,7 +126,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else {
@@ -141,7 +141,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else /* there are NULLs in the inputColVector */ {
@@ -158,7 +158,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
@@ -170,7 +170,7 @@ public class <ClassName> extends VectorExpression {
scratchIntervalYearMonth1.set((int) vector1[i]);
dtm.<OperatorMethod>(
scratchIntervalYearMonth1, value, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
outputIsNull[i] = false;
} else {
outputIsNull[i] = true;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticTimestampScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticTimestampScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticTimestampScalar.txt
index 71e142c..c297116 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticTimestampScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthColumnArithmeticTimestampScalar.txt
@@ -182,7 +182,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return getColumnParamString(0, colNum) + ", val " + TimestampUtils.timestampScalarTypeToString(value);
+ return getColumnParamString(0, colNum) + ", val " + value.toString();
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthScalarArithmeticDateColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthScalarArithmeticDateColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthScalarArithmeticDateColumn.txt
index 27f2fcf..34d516e 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthScalarArithmeticDateColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/IntervalYearMonthScalarArithmeticDateColumn.txt
@@ -37,8 +37,8 @@ import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template DateTimeScalarArithmeticIntervalYearMonthColumn.txt.
@@ -107,10 +107,10 @@ public class <ClassName> extends VectorExpression {
if (inputColVector2.isRepeating) {
if (inputColVector2.noNulls || !inputIsNull[0]) {
outputIsNull[0] = false;
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[0]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[0]));
dtm.<OperatorMethod>(
value, scratchDate2, outputDate);
- outputVector[0] = DateWritableV2.dateToDays(outputDate);
+ outputVector[0] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[0] = true;
outputColVector.noNulls = false;
@@ -129,18 +129,18 @@ public class <ClassName> extends VectorExpression {
for(int j = 0; j != n; j++) {
final int i = sel[j];
outputIsNull[i] = false;
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
} else {
for(int j = 0; j != n; j++) {
final int i = sel[j];
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else {
@@ -152,10 +152,10 @@ public class <ClassName> extends VectorExpression {
outputColVector.noNulls = true;
}
for(int i = 0; i != n; i++) {
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
}
}
} else /* there are NULLs in the inputColVector */ {
@@ -169,10 +169,10 @@ public class <ClassName> extends VectorExpression {
int i = sel[j];
if (!inputIsNull[i]) {
outputIsNull[i] = false;
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
@@ -182,10 +182,10 @@ public class <ClassName> extends VectorExpression {
for(int i = 0; i != n; i++) {
if (!inputIsNull[i]) {
outputIsNull[i] = false;
- scratchDate2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchDate2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value, scratchDate2, outputDate);
- outputVector[i] = DateWritableV2.dateToDays(outputDate);
+ outputVector[i] = DateWritable.dateToDays(outputDate);
} else {
outputIsNull[i] = true;
outputColVector.noNulls = false;
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/LongDoubleColumnCompareTimestampScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/LongDoubleColumnCompareTimestampScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/LongDoubleColumnCompareTimestampScalar.txt
index 397fca5..19263d9 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/LongDoubleColumnCompareTimestampScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/LongDoubleColumnCompareTimestampScalar.txt
@@ -157,7 +157,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return getColumnParamString(0, colNum) + ", val " + TimestampUtils.timestampScalarTypeToString(value);
+ return getColumnParamString(0, colNum) + ", val " + value;
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateColumn.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateColumn.txt b/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateColumn.txt
index 8e44c92..45f6408 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateColumn.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateColumn.txt
@@ -27,8 +27,8 @@ import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template TimestampColumnArithmeticDateColumn.txt, which covers binary arithmetic
@@ -96,7 +96,7 @@ public class <ClassName> extends VectorExpression {
* conditional checks in the inner loop.
*/
if (inputColVector1.isRepeating && inputColVector2.isRepeating) {
- scratchTimestamp2.setTime(DateWritableV2.daysToMillis((int) vector2[0]));
+ scratchTimestamp2.setTime(DateWritable.daysToMillis((int) vector2[0]));
dtm.<OperatorMethod>(
inputColVector1.asScratch<CamelOperandType1>(0), scratchTimestamp2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(0);
@@ -105,21 +105,21 @@ public class <ClassName> extends VectorExpression {
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchTimestamp2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchTimestamp2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value1, scratchTimestamp2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
}
} else {
for(int i = 0; i != n; i++) {
- scratchTimestamp2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchTimestamp2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
value1, scratchTimestamp2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
}
}
} else if (inputColVector2.isRepeating) {
- scratchTimestamp2.setTime(DateWritableV2.daysToMillis((int) vector2[0]));
+ scratchTimestamp2.setTime(DateWritable.daysToMillis((int) vector2[0]));
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
@@ -138,14 +138,14 @@ public class <ClassName> extends VectorExpression {
if (batch.selectedInUse) {
for(int j = 0; j != n; j++) {
int i = sel[j];
- scratchTimestamp2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchTimestamp2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
inputColVector1.asScratch<CamelOperandType1>(i), scratchTimestamp2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
}
} else {
for(int i = 0; i != n; i++) {
- scratchTimestamp2.setTime(DateWritableV2.daysToMillis((int) vector2[i]));
+ scratchTimestamp2.setTime(DateWritable.daysToMillis((int) vector2[i]));
dtm.<OperatorMethod>(
inputColVector1.asScratch<CamelOperandType1>(i), scratchTimestamp2, outputColVector.getScratch<CamelReturnType>());
outputColVector.setFromScratch<CamelReturnType>(i);
http://git-wip-us.apache.org/repos/asf/hive/blob/33088de0/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateScalar.txt
----------------------------------------------------------------------
diff --git a/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateScalar.txt b/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateScalar.txt
index cb897e4..0bbdce7 100644
--- a/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateScalar.txt
+++ b/ql/src/gen/vectorization/ExpressionTemplates/TimestampColumnArithmeticDateScalar.txt
@@ -29,8 +29,8 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.NullUtil;
import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
import org.apache.hadoop.hive.ql.exec.vector.*;
import org.apache.hadoop.hive.ql.util.DateTimeMath;
+import org.apache.hadoop.hive.serde2.io.DateWritable;
import org.apache.hadoop.hive.ql.metadata.HiveException;
-import org.apache.hadoop.hive.serde2.io.DateWritableV2;
/**
* Generated from template TimestampColumnArithmeticDateScalar.txt, which covers binary arithmetic
@@ -49,7 +49,7 @@ public class <ClassName> extends VectorExpression {
super(outputColumnNum);
this.colNum = colNum;
this.value = new Timestamp(0);
- this.value.setTime(DateWritableV2.daysToMillis((int) value));
+ this.value.setTime(DateWritable.daysToMillis((int) value));
}
public <ClassName>() {
@@ -175,7 +175,7 @@ public class <ClassName> extends VectorExpression {
@Override
public String vectorExpressionParameters() {
- return getColumnParamString(0, colNum) + ", val " + org.apache.hadoop.hive.common.type.Date.ofEpochMilli(value.getTime());
+ return getColumnParamString(0, colNum) + ", val " + value.toString();
}
@Override