You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2013/07/31 00:22:46 UTC
svn commit: r1508669 [36/39] - in /hive/branches/vectorization: ./
common/src/java/org/apache/hadoop/hive/conf/ conf/
contrib/src/test/results/clientpositive/ data/files/ eclipse-templates/
hcatalog/build-support/ant/ hcatalog/core/src/main/java/org/ap...
Modified: hive/branches/vectorization/ql/src/test/results/compiler/plan/udf6.q.xml
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/compiler/plan/udf6.q.xml?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/compiler/plan/udf6.q.xml (original)
+++ hive/branches/vectorization/ql/src/test/results/compiler/plan/udf6.q.xml Tue Jul 30 22:22:35 2013
@@ -6,158 +6,161 @@
</void>
<void property="work">
<object class="org.apache.hadoop.hive.ql.plan.MapredWork">
- <void property="aliasToPartnInfo">
- <void method="put">
- <string>src</string>
- <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="partSpec">
- <object class="java.util.LinkedHashMap"/>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void property="tableDesc">
- <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>numFiles</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>rawDataSize</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numRows</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numPartitions</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>totalSize</string>
- <string>5812</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- </object>
- </void>
- </object>
- </void>
+ <void property="finalMapRed">
+ <boolean>true</boolean>
</void>
- <void property="aliasToWork">
- <object class="java.util.LinkedHashMap">
+ <void property="mapWork">
+ <void property="aliasToPartnInfo">
+ <void method="put">
+ <string>src</string>
+ <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="partSpec">
+ <object class="java.util.LinkedHashMap"/>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void property="tableDesc">
+ <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>numFiles</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>rawDataSize</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numRows</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numPartitions</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>totalSize</string>
+ <string>5812</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </void>
+ <void property="aliasToWork">
<void method="put">
<string>src</string>
<object id="TableScanOperator0" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
@@ -587,16 +590,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="finalMapRed">
- <boolean>true</boolean>
- </void>
- <void property="nameToSplitSample">
- <object class="java.util.HashMap"/>
- </void>
- <void property="pathToAliases">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToAliases">
<void method="put">
#### A masked pattern was here ####
<object class="java.util.ArrayList">
@@ -605,10 +600,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="pathToPartitionInfo">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToPartitionInfo">
<void method="put">
#### A masked pattern was here ####
<object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
@@ -784,10 +777,7 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="tagToValueDesc">
- <object class="java.util.ArrayList"/>
+ </void>
</void>
</object>
</void>
Modified: hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_case.q.xml
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_case.q.xml?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_case.q.xml (original)
+++ hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_case.q.xml Tue Jul 30 22:22:35 2013
@@ -6,158 +6,161 @@
</void>
<void property="work">
<object class="org.apache.hadoop.hive.ql.plan.MapredWork">
- <void property="aliasToPartnInfo">
- <void method="put">
- <string>src</string>
- <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="partSpec">
- <object class="java.util.LinkedHashMap"/>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void property="tableDesc">
- <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>numFiles</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>rawDataSize</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numRows</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numPartitions</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>totalSize</string>
- <string>5812</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- </object>
- </void>
- </object>
- </void>
+ <void property="finalMapRed">
+ <boolean>true</boolean>
</void>
- <void property="aliasToWork">
- <object class="java.util.LinkedHashMap">
+ <void property="mapWork">
+ <void property="aliasToPartnInfo">
+ <void method="put">
+ <string>src</string>
+ <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="partSpec">
+ <object class="java.util.LinkedHashMap"/>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void property="tableDesc">
+ <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>numFiles</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>rawDataSize</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numRows</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numPartitions</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>totalSize</string>
+ <string>5812</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </void>
+ <void property="aliasToWork">
<void method="put">
<string>src</string>
<object id="TableScanOperator0" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
@@ -681,16 +684,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="finalMapRed">
- <boolean>true</boolean>
- </void>
- <void property="nameToSplitSample">
- <object class="java.util.HashMap"/>
- </void>
- <void property="pathToAliases">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToAliases">
<void method="put">
#### A masked pattern was here ####
<object class="java.util.ArrayList">
@@ -699,10 +694,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="pathToPartitionInfo">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToPartitionInfo">
<void method="put">
#### A masked pattern was here ####
<object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
@@ -878,10 +871,7 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="tagToValueDesc">
- <object class="java.util.ArrayList"/>
+ </void>
</void>
</object>
</void>
Modified: hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_when.q.xml
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_when.q.xml?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_when.q.xml (original)
+++ hive/branches/vectorization/ql/src/test/results/compiler/plan/udf_when.q.xml Tue Jul 30 22:22:35 2013
@@ -6,158 +6,161 @@
</void>
<void property="work">
<object class="org.apache.hadoop.hive.ql.plan.MapredWork">
- <void property="aliasToPartnInfo">
- <void method="put">
- <string>src</string>
- <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="partSpec">
- <object class="java.util.LinkedHashMap"/>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void property="tableDesc">
- <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>numFiles</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>rawDataSize</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numRows</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numPartitions</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>totalSize</string>
- <string>5812</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- </object>
- </void>
- </object>
- </void>
+ <void property="finalMapRed">
+ <boolean>true</boolean>
</void>
- <void property="aliasToWork">
- <object class="java.util.LinkedHashMap">
+ <void property="mapWork">
+ <void property="aliasToPartnInfo">
+ <void method="put">
+ <string>src</string>
+ <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="partSpec">
+ <object class="java.util.LinkedHashMap"/>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void property="tableDesc">
+ <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>numFiles</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>rawDataSize</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numRows</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numPartitions</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>totalSize</string>
+ <string>5812</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </void>
+ <void property="aliasToWork">
<void method="put">
<string>src</string>
<object id="TableScanOperator0" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
@@ -761,16 +764,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="finalMapRed">
- <boolean>true</boolean>
- </void>
- <void property="nameToSplitSample">
- <object class="java.util.HashMap"/>
- </void>
- <void property="pathToAliases">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToAliases">
<void method="put">
#### A masked pattern was here ####
<object class="java.util.ArrayList">
@@ -779,10 +774,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="pathToPartitionInfo">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToPartitionInfo">
<void method="put">
#### A masked pattern was here ####
<object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
@@ -958,10 +951,7 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="tagToValueDesc">
- <object class="java.util.ArrayList"/>
+ </void>
</void>
</object>
</void>
Modified: hive/branches/vectorization/ql/src/test/results/compiler/plan/union.q.xml
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/ql/src/test/results/compiler/plan/union.q.xml?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/ql/src/test/results/compiler/plan/union.q.xml (original)
+++ hive/branches/vectorization/ql/src/test/results/compiler/plan/union.q.xml Tue Jul 30 22:22:35 2013
@@ -38,8 +38,8 @@
</void>
<void property="work">
<object id="MapredWork0" class="org.apache.hadoop.hive.ql.plan.MapredWork">
- <void property="aliasToWork">
- <object class="java.util.LinkedHashMap">
+ <void property="mapWork">
+ <void property="aliasToWork">
<void method="put">
#### A masked pattern was here ####
<object id="TableScanOperator0" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
@@ -181,19 +181,14 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="inputformat">
- <string>org.apache.hadoop.hive.ql.io.CombineHiveInputFormat</string>
- </void>
- <void property="mapperCannotSpanPartns">
- <boolean>true</boolean>
- </void>
- <void property="numReduceTasks">
- <int>0</int>
- </void>
- <void property="pathToAliases">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="inputformat">
+ <string>org.apache.hadoop.hive.ql.io.CombineHiveInputFormat</string>
+ </void>
+ <void property="mapperCannotSpanPartns">
+ <boolean>true</boolean>
+ </void>
+ <void property="pathToAliases">
<void method="put">
#### A masked pattern was here ####
<object class="java.util.ArrayList">
@@ -202,10 +197,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="pathToPartitionInfo">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToPartitionInfo">
<void method="put">
#### A masked pattern was here ####
<object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
@@ -255,10 +248,7 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="tagToValueDesc">
- <object class="java.util.ArrayList"/>
+ </void>
</void>
</object>
</void>
@@ -404,306 +394,309 @@
</void>
<void property="work">
<object class="org.apache.hadoop.hive.ql.plan.MapredWork">
- <void property="aliasToPartnInfo">
- <void method="put">
- <string>null-subquery1:unioninput-subquery1:src</string>
- <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="partSpec">
- <object class="java.util.LinkedHashMap"/>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void property="tableDesc">
- <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>numFiles</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>rawDataSize</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numRows</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numPartitions</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>totalSize</string>
- <string>5812</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- </object>
- </void>
- </object>
- </void>
- <void method="put">
- <string>null-subquery2:unioninput-subquery2:src</string>
- <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="partSpec">
- <object class="java.util.LinkedHashMap"/>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void property="tableDesc">
- <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>default.src</string>
- </void>
- <void method="put">
- <string>numFiles</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct src { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>rawDataSize</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numRows</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>numPartitions</string>
- <string>0</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>totalSize</string>
- <string>5812</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- #### A masked pattern was here ####
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- #### A masked pattern was here ####
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- </object>
- </void>
- </object>
- </void>
+ <void property="finalMapRed">
+ <boolean>true</boolean>
</void>
- <void property="aliasToWork">
- <object class="java.util.LinkedHashMap">
+ <void property="mapWork">
+ <void property="aliasToPartnInfo">
+ <void method="put">
+ <string>null-subquery1:unioninput-subquery1:src</string>
+ <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="partSpec">
+ <object class="java.util.LinkedHashMap"/>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void property="tableDesc">
+ <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>numFiles</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>rawDataSize</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numRows</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numPartitions</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>totalSize</string>
+ <string>5812</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void method="put">
+ <string>null-subquery2:unioninput-subquery2:src</string>
+ <object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="partSpec">
+ <object class="java.util.LinkedHashMap"/>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void property="tableDesc">
+ <object class="org.apache.hadoop.hive.ql.plan.TableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>default.src</string>
+ </void>
+ <void method="put">
+ <string>numFiles</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct src { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>rawDataSize</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numRows</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>numPartitions</string>
+ <string>0</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>totalSize</string>
+ <string>5812</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ #### A masked pattern was here ####
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ #### A masked pattern was here ####
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </void>
+ <void property="aliasToWork">
<void method="put">
<string>null-subquery1:unioninput-subquery1:src</string>
<object id="TableScanOperator1" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
@@ -1063,12 +1056,12 @@
<void property="typeInfo">
<object id="PrimitiveTypeInfo1" class="org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo">
<void property="typeName">
- <string>double</string>
+ <string>int</string>
</void>
</object>
</void>
<void property="value">
- <double>100.0</double>
+ <int>100</int>
</void>
</object>
</void>
@@ -1470,7 +1463,7 @@
<object idref="PrimitiveTypeInfo1"/>
</void>
<void property="value">
- <double>100.0</double>
+ <int>100</int>
</void>
</object>
</void>
@@ -1641,16 +1634,8 @@
<string>null-subquery2:unioninput-subquery2:src</string>
<object idref="TableScanOperator2"/>
</void>
- </object>
- </void>
- <void property="finalMapRed">
- <boolean>true</boolean>
- </void>
- <void property="nameToSplitSample">
- <object class="java.util.HashMap"/>
- </void>
- <void property="pathToAliases">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToAliases">
<void method="put">
#### A masked pattern was here ####
<object class="java.util.ArrayList">
@@ -1662,10 +1647,8 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="pathToPartitionInfo">
- <object class="java.util.LinkedHashMap">
+ </void>
+ <void property="pathToPartitionInfo">
<void method="put">
#### A masked pattern was here ####
<object class="org.apache.hadoop.hive.ql.plan.PartitionDesc">
@@ -1841,10 +1824,7 @@
</void>
</object>
</void>
- </object>
- </void>
- <void property="tagToValueDesc">
- <object class="java.util.ArrayList"/>
+ </void>
</void>
</object>
</void>
Modified: hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java (original)
+++ hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java Tue Jul 30 22:22:35 2013
@@ -17,6 +17,8 @@
*/
package org.apache.hadoop.hive.serde2;
+import java.sql.Date;
+import java.sql.Timestamp;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
@@ -134,6 +136,10 @@ public class RegexSerDe extends Abstract
columnOIs.add(PrimitiveObjectInspectorFactory.javaDoubleObjectInspector);
} else if (typeName.equals(serdeConstants.BOOLEAN_TYPE_NAME)) {
columnOIs.add(PrimitiveObjectInspectorFactory.javaBooleanObjectInspector);
+ } else if (typeName.equals(serdeConstants.TIMESTAMP_TYPE_NAME)) {
+ columnOIs.add(PrimitiveObjectInspectorFactory.javaTimestampObjectInspector);
+ } else if (typeName.equals(serdeConstants.DATE_TYPE_NAME)) {
+ columnOIs.add(PrimitiveObjectInspectorFactory.javaDateObjectInspector);
} else if (typeName.equals(serdeConstants.DECIMAL_TYPE_NAME)) {
columnOIs.add(PrimitiveObjectInspectorFactory.javaHiveDecimalObjectInspector);
} else {
@@ -229,6 +235,14 @@ public class RegexSerDe extends Abstract
Boolean b;
b = Boolean.valueOf(t);
row.set(c, b);
+ } else if (typeName.equals(serdeConstants.TIMESTAMP_TYPE_NAME)) {
+ Timestamp ts;
+ ts = Timestamp.valueOf(t);
+ row.set(c, ts);
+ } else if (typeName.equals(serdeConstants.DATE_TYPE_NAME)) {
+ Date d;
+ d = Date.valueOf(t);
+ row.set(c, d);
} else if (typeName.equals(serdeConstants.DECIMAL_TYPE_NAME)) {
HiveDecimal bd;
bd = new HiveDecimal(t);
Modified: hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java (original)
+++ hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java Tue Jul 30 22:22:35 2013
@@ -44,6 +44,7 @@ import org.apache.hadoop.hive.serde2.obj
import org.apache.hadoop.hive.serde2.objectinspector.primitive.LongObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.ShortObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.StringObjectInspector;
+import org.apache.hadoop.hive.serde2.objectinspector.primitive.DateObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.TimestampObjectInspector;
import org.apache.hadoop.io.BytesWritable;
import org.apache.hadoop.io.Text;
@@ -283,6 +284,13 @@ public final class SerDeUtils {
sb.append('"');
break;
}
+ case DATE: {
+ sb.append('"');
+ sb.append(((DateObjectInspector) poi)
+ .getPrimitiveWritableObject(o));
+ sb.append('"');
+ break;
+ }
case TIMESTAMP: {
sb.append('"');
sb.append(((TimestampObjectInspector) poi)
Modified: hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java
URL: http://svn.apache.org/viewvc/hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java?rev=1508669&r1=1508668&r2=1508669&view=diff
==============================================================================
--- hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java (original)
+++ hive/branches/vectorization/serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java Tue Jul 30 22:22:35 2013
@@ -17,8 +17,19 @@
*/
package org.apache.hadoop.hive.serde2.avro;
+import java.io.ByteArrayInputStream;
+import java.io.ByteArrayOutputStream;
+import java.io.IOException;
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
import org.apache.avro.Schema;
+import org.apache.avro.Schema.Type;
import org.apache.avro.generic.GenericData;
+import org.apache.avro.generic.GenericData.Fixed;
import org.apache.avro.generic.GenericDatumReader;
import org.apache.avro.generic.GenericDatumWriter;
import org.apache.avro.generic.GenericRecord;
@@ -32,21 +43,12 @@ import org.apache.commons.logging.LogFac
import org.apache.hadoop.hive.serde2.objectinspector.StandardUnionObjectInspector;
import org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.MapTypeInfo;
+import org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.StructTypeInfo;
import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
-import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory;
import org.apache.hadoop.hive.serde2.typeinfo.UnionTypeInfo;
import org.apache.hadoop.io.Writable;
-import java.io.ByteArrayInputStream;
-import java.io.ByteArrayOutputStream;
-import java.io.IOException;
-import java.nio.ByteBuffer;
-import java.util.ArrayList;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-
class AvroDeserializer {
private static final Log LOG = LogFactory.getLog(AvroDeserializer.class);
/**
@@ -62,7 +64,7 @@ class AvroDeserializer {
private final ByteArrayOutputStream baos = new ByteArrayOutputStream();
private final GenericDatumWriter<GenericRecord> gdw = new GenericDatumWriter<GenericRecord>();
private BinaryDecoder binaryDecoder = null;
- private InstanceCache<ReaderWriterSchemaPair, GenericDatumReader<GenericRecord>> gdrCache
+ private final InstanceCache<ReaderWriterSchemaPair, GenericDatumReader<GenericRecord>> gdrCache
= new InstanceCache<ReaderWriterSchemaPair, GenericDatumReader<GenericRecord>>() {
@Override
protected GenericDatumReader<GenericRecord> makeInstance(ReaderWriterSchemaPair hv) {
@@ -112,13 +114,15 @@ class AvroDeserializer {
*/
public Object deserialize(List<String> columnNames, List<TypeInfo> columnTypes,
Writable writable, Schema readerSchema) throws AvroSerdeException {
- if(!(writable instanceof AvroGenericRecordWritable))
+ if(!(writable instanceof AvroGenericRecordWritable)) {
throw new AvroSerdeException("Expecting a AvroGenericRecordWritable");
+ }
- if(row == null || row.size() != columnNames.size())
+ if(row == null || row.size() != columnNames.size()) {
row = new ArrayList<Object>(columnNames.size());
- else
+ } else {
row.clear();
+ }
AvroGenericRecordWritable recordWritable = (AvroGenericRecordWritable) writable;
GenericRecord r = recordWritable.getRecord();
@@ -127,7 +131,9 @@ class AvroDeserializer {
if(!r.getSchema().equals(readerSchema)) {
LOG.warn("Received different schemas. Have to re-encode: " +
r.getSchema().toString(false));
- if(reEncoder == null) reEncoder = new SchemaReEncoder();
+ if(reEncoder == null) {
+ reEncoder = new SchemaReEncoder();
+ }
r = reEncoder.reencode(r, readerSchema);
}
@@ -156,25 +162,49 @@ class AvroDeserializer {
// Klaxon! Klaxon! Klaxon!
// Avro requires NULLable types to be defined as unions of some type T
// and NULL. This is annoying and we're going to hide it from the user.
- if(AvroSerdeUtils.isNullableType(recordSchema))
+ if(AvroSerdeUtils.isNullableType(recordSchema)) {
return deserializeNullableUnion(datum, recordSchema, columnType);
+ }
- if(columnType == TypeInfoFactory.stringTypeInfo)
- return datum.toString(); // To workaround AvroUTF8
- // This also gets us around the Enum issue since we just take the value
- // and convert it to a string. Yay!
switch(columnType.getCategory()) {
case STRUCT:
return deserializeStruct((GenericData.Record) datum, (StructTypeInfo) columnType);
- case UNION:
+ case UNION:
return deserializeUnion(datum, recordSchema, (UnionTypeInfo) columnType);
case LIST:
return deserializeList(datum, recordSchema, (ListTypeInfo) columnType);
case MAP:
return deserializeMap(datum, recordSchema, (MapTypeInfo) columnType);
+ case PRIMITIVE:
+ return deserializePrimitive(datum, recordSchema, (PrimitiveTypeInfo) columnType);
default:
- return datum; // Simple type.
+ throw new AvroSerdeException("Unknown TypeInfo: " + columnType.getCategory());
+ }
+ }
+
+ private Object deserializePrimitive(Object datum, Schema recordSchema,
+ PrimitiveTypeInfo columnType) throws AvroSerdeException {
+ switch (columnType.getPrimitiveCategory()){
+ case STRING:
+ return datum.toString(); // To workaround AvroUTF8
+ // This also gets us around the Enum issue since we just take the value
+ // and convert it to a string. Yay!
+ case BINARY:
+ if (recordSchema.getType() == Type.FIXED){
+ Fixed fixed = (Fixed) datum;
+ return fixed.bytes();
+ } else if (recordSchema.getType() == Type.BYTES){
+ ByteBuffer bb = (ByteBuffer) datum;
+ bb.rewind();
+ byte[] result = new byte[bb.limit()];
+ bb.get(result);
+ return result;
+ } else {
+ throw new AvroSerdeException("Unexpected Avro schema for Binary TypeInfo: " + recordSchema.getType());
+ }
+ default:
+ return datum;
}
}
@@ -186,8 +216,9 @@ class AvroDeserializer {
TypeInfo columnType) throws AvroSerdeException {
int tag = GenericData.get().resolveUnion(recordSchema, datum); // Determine index of value
Schema schema = recordSchema.getTypes().get(tag);
- if(schema.getType().equals(Schema.Type.NULL))
+ if(schema.getType().equals(Schema.Type.NULL)) {
return null;
+ }
return worker(datum, schema, SchemaToTypeInfo.generateTypeInfo(schema));
}