You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by na...@apache.org on 2010/01/16 07:46:07 UTC
svn commit: r899891 [20/31] - in /hadoop/hive/trunk: ./
common/src/java/org/apache/hadoop/hive/conf/ conf/
ql/src/java/org/apache/hadoop/hive/ql/
ql/src/java/org/apache/hadoop/hive/ql/exec/
ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/ ql/src...
Modified: hadoop/hive/trunk/ql/src/test/results/compiler/plan/input6.q.xml
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/ql/src/test/results/compiler/plan/input6.q.xml?rev=899891&r1=899890&r2=899891&view=diff
==============================================================================
--- hadoop/hive/trunk/ql/src/test/results/compiler/plan/input6.q.xml (original)
+++ hadoop/hive/trunk/ql/src/test/results/compiler/plan/input6.q.xml Sat Jan 16 06:44:01 2010
@@ -4,258 +4,501 @@
<void property="childTasks">
<object class="java.util.ArrayList">
<void method="add">
- <object id="ConditionalTask0" class="org.apache.hadoop.hive.ql.exec.ConditionalTask">
- <void property="childTasks">
- <object class="java.util.ArrayList">
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.MoveTask">
- <void property="id">
- <string>Stage-1</string>
- </void>
- <void property="parentTasks">
- <object class="java.util.ArrayList">
- <void method="add">
- <object idref="ConditionalTask0"/>
- </void>
- </object>
- </void>
- <void property="work">
- <object class="org.apache.hadoop.hive.ql.plan.moveWork">
- <void property="loadTableWork">
- <object class="org.apache.hadoop.hive.ql.plan.loadTableDesc">
- <void property="partitionSpec">
- <object class="java.util.HashMap"/>
- </void>
- <void property="replace">
- <boolean>true</boolean>
- </void>
- <void property="sourceDir">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1704458189/10000</string>
- </void>
- <void property="table">
- <object id="tableDesc0" class="org.apache.hadoop.hive.ql.plan.tableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object id="Properties0" class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>dest1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct dest1 { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1</string>
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- <string>1263415292</string>
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- </object>
- </void>
- <void property="tmpDir">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1704458189/10001</string>
- </void>
- </object>
- </void>
- </object>
- </void>
- </object>
- </void>
- </object>
- </void>
+ <object class="org.apache.hadoop.hive.ql.exec.ConditionalTask">
<void property="id">
<string>Stage-5</string>
</void>
<void property="listTasks">
<object id="ArrayList0" class="java.util.ArrayList">
<void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.MoveTask">
- <void property="id">
- <string>Stage-4</string>
- </void>
- <void property="work">
- <object id="moveWork0" class="org.apache.hadoop.hive.ql.plan.moveWork">
- <void property="loadFileWork">
- <object class="org.apache.hadoop.hive.ql.plan.loadFileDesc">
- <void property="isDfsDir">
- <boolean>true</boolean>
- </void>
- <void property="sourceDir">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
- </void>
- <void property="targetDir">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1704458189/10000</string>
+ <object id="MoveTask0" class="org.apache.hadoop.hive.ql.exec.MoveTask">
+ <void property="childTasks">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object id="MoveTask1" class="org.apache.hadoop.hive.ql.exec.MoveTask">
+ <void property="id">
+ <string>Stage-1</string>
</void>
- </object>
- </void>
- </object>
- </void>
- </object>
- </void>
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.MapRedTask">
- <void property="id">
- <string>Stage-3</string>
- </void>
- <void property="work">
- <object id="mapredWork0" class="org.apache.hadoop.hive.ql.plan.mapredWork">
- <void property="aliasToWork">
- <object class="java.util.LinkedHashMap">
- <void method="put">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
- <object id="TableScanOperator0" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
- <void property="childOperators">
- <object class="java.util.ArrayList">
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.ReduceSinkOperator">
- <void property="conf">
- <object class="org.apache.hadoop.hive.ql.plan.reduceSinkDesc">
- <void property="keyCols">
- <object class="java.util.ArrayList"/>
+ <void property="parentTasks">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object idref="MoveTask0"/>
+ </void>
+ <void method="add">
+ <object id="MapRedTask1" class="org.apache.hadoop.hive.ql.exec.MapRedTask">
+ <void property="childTasks">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object idref="MoveTask1"/>
+ </void>
+ </object>
+ </void>
+ <void property="id">
+ <string>Stage-3</string>
+ </void>
+ <void property="work">
+ <object id="mapredWork0" class="org.apache.hadoop.hive.ql.plan.mapredWork">
+ <void property="aliasToWork">
+ <object class="java.util.LinkedHashMap">
+ <void method="put">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
+ <object id="TableScanOperator0" class="org.apache.hadoop.hive.ql.exec.TableScanOperator">
+ <void property="childOperators">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.exec.ReduceSinkOperator">
+ <void property="conf">
+ <object class="org.apache.hadoop.hive.ql.plan.reduceSinkDesc">
+ <void property="keyCols">
+ <object class="java.util.ArrayList"/>
+ </void>
+ <void property="keySerializeInfo">
+ <object id="tableDesc0" class="org.apache.hadoop.hive.ql.plan.tableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.SequenceFileInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>columns</string>
+ <string></string>
+ </void>
+ <void method="put">
+ <string>serialization.sort.order</string>
+ <string></string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string></string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void property="numReducers">
+ <int>-1</int>
+ </void>
+ <void property="outputKeyColumnNames">
+ <object class="java.util.ArrayList"/>
+ </void>
+ <void property="outputValueColumnNames">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <string>_col0</string>
+ </void>
+ <void method="add">
+ <string>_col1</string>
+ </void>
+ </object>
+ </void>
+ <void property="partitionCols">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.plan.exprNodeGenericFuncDesc">
+ <void property="childExprs">
+ <object class="java.util.ArrayList"/>
+ </void>
+ <void property="genericUDF">
+ <object class="org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge">
+ <void property="udfClass">
+ <class>org.apache.hadoop.hive.ql.udf.UDFRand</class>
+ </void>
+ <void property="udfName">
+ <string>rand</string>
+ </void>
+ </object>
+ </void>
+ <void property="typeInfo">
+ <object class="org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo">
+ <void property="typeName">
+ <string>double</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void property="tag">
+ <int>-1</int>
+ </void>
+ <void property="valueCols">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc">
+ <void property="column">
+ <string>key</string>
+ </void>
+ <void property="tabAlias">
+ <string></string>
+ </void>
+ <void property="typeInfo">
+ <object id="PrimitiveTypeInfo0" class="org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo">
+ <void property="typeName">
+ <string>string</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc">
+ <void property="column">
+ <string>value</string>
+ </void>
+ <void property="tabAlias">
+ <string></string>
+ </void>
+ <void property="typeInfo">
+ <object idref="PrimitiveTypeInfo0"/>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void property="valueSerializeInfo">
+ <object id="tableDesc1" class="org.apache.hadoop.hive.ql.plan.tableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.SequenceFileInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>columns</string>
+ <string>_col0,_col1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string,string</string>
+ </void>
+ <void method="put">
+ <string>escape.delim</string>
+ <string>\</string>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void property="counterNames">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <string>CNTR_NAME_RS_6_NUM_INPUT_ROWS</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_RS_6_NUM_OUTPUT_ROWS</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_RS_6_TIME_TAKEN</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_RS_6_FATAL_ERROR</string>
+ </void>
+ </object>
+ </void>
+ <void property="operatorId">
+ <string>RS_6</string>
+ </void>
+ <void property="parentOperators">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object idref="TableScanOperator0"/>
+ </void>
+ </object>
+ </void>
+ <void property="schema">
+ <object id="RowSchema0" class="org.apache.hadoop.hive.ql.exec.RowSchema">
+ <void property="signature">
+ <object class="java.util.Vector">
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
+ <void property="internalName">
+ <string>key</string>
+ </void>
+ <void property="type">
+ <object idref="PrimitiveTypeInfo0"/>
+ </void>
+ </object>
+ </void>
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
+ <void property="internalName">
+ <string>value</string>
+ </void>
+ <void property="type">
+ <object idref="PrimitiveTypeInfo0"/>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void property="counterNames">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <string>CNTR_NAME_TS_5_NUM_INPUT_ROWS</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_TS_5_NUM_OUTPUT_ROWS</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_TS_5_TIME_TAKEN</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_TS_5_FATAL_ERROR</string>
+ </void>
+ </object>
+ </void>
+ <void property="operatorId">
+ <string>TS_5</string>
+ </void>
+ <void property="schema">
+ <object idref="RowSchema0"/>
+ </void>
+ </object>
+ </void>
+ </object>
+ </void>
+ <void property="keyDesc">
+ <object idref="tableDesc0"/>
+ </void>
+ <void property="numReduceTasks">
+ <int>-1</int>
+ </void>
+ <void property="pathToAliases">
+ <object class="java.util.LinkedHashMap">
+ <void method="put">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
+ </void>
+ </object>
</void>
- <void property="keySerializeInfo">
- <object id="tableDesc1" class="org.apache.hadoop.hive.ql.plan.tableDesc">
+ </object>
+ </void>
+ <void property="pathToPartitionInfo">
+ <object class="java.util.LinkedHashMap">
+ <void method="put">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
+ <object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe</class>
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
</void>
<void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.SequenceFileInputFormat</class>
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
</void>
<void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat</class>
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
</void>
<void property="properties">
- <object class="java.util.Properties">
+ <object id="Properties0" class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>dest1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct dest1 { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
<void method="put">
<string>columns</string>
- <string></string>
+ <string>key,value</string>
</void>
<void method="put">
- <string>serialization.sort.order</string>
- <string></string>
+ <string>bucket_count</string>
+ <string>-1</string>
</void>
<void method="put">
- <string>columns.types</string>
- <string></string>
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/test/data/warehouse/dest1</string>
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ <string>1263595117</string>
</void>
</object>
</void>
- </object>
- </void>
- <void property="numReducers">
- <int>-1</int>
- </void>
- <void property="outputKeyColumnNames">
- <object class="java.util.ArrayList"/>
- </void>
- <void property="outputValueColumnNames">
- <object class="java.util.ArrayList">
- <void method="add">
- <string>_col0</string>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
</void>
- <void method="add">
- <string>_col1</string>
+ <void property="tableDesc">
+ <object id="tableDesc2" class="org.apache.hadoop.hive.ql.plan.tableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object idref="Properties0"/>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
</void>
</object>
</void>
- <void property="partitionCols">
+ </object>
+ </void>
+ <void property="reducer">
+ <object id="ExtractOperator0" class="org.apache.hadoop.hive.ql.exec.ExtractOperator">
+ <void property="childOperators">
<object class="java.util.ArrayList">
<void method="add">
- <object class="org.apache.hadoop.hive.ql.plan.exprNodeGenericFuncDesc">
- <void property="childExprs">
- <object class="java.util.ArrayList"/>
- </void>
- <void property="genericUDF">
- <object class="org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge">
- <void property="udfClass">
- <class>org.apache.hadoop.hive.ql.udf.UDFRand</class>
+ <object class="org.apache.hadoop.hive.ql.exec.FileSinkOperator">
+ <void property="conf">
+ <object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc">
+ <void property="dirName">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/200354901/10000</string>
</void>
- <void property="udfName">
- <string>rand</string>
+ <void property="tableInfo">
+ <object class="org.apache.hadoop.hive.ql.plan.tableDesc">
+ <void property="deserializerClass">
+ <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
+ </void>
+ <void property="inputFileFormatClass">
+ <class>org.apache.hadoop.mapred.TextInputFormat</class>
+ </void>
+ <void property="outputFileFormatClass">
+ <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
+ </void>
+ <void property="properties">
+ <object class="java.util.Properties">
+ <void method="put">
+ <string>name</string>
+ <string>dest1</string>
+ </void>
+ <void method="put">
+ <string>columns.types</string>
+ <string>string:string</string>
+ </void>
+ <void method="put">
+ <string>serialization.ddl</string>
+ <string>struct dest1 { string key, string value}</string>
+ </void>
+ <void method="put">
+ <string>serialization.format</string>
+ <string>1</string>
+ </void>
+ <void method="put">
+ <string>columns</string>
+ <string>key,value</string>
+ </void>
+ <void method="put">
+ <string>bucket_count</string>
+ <string>-1</string>
+ </void>
+ <void method="put">
+ <string>serialization.lib</string>
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ <void method="put">
+ <string>file.inputformat</string>
+ <string>org.apache.hadoop.mapred.TextInputFormat</string>
+ </void>
+ <void method="put">
+ <string>location</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/test/data/warehouse/dest1</string>
+ </void>
+ <void method="put">
+ <string>file.outputformat</string>
+ <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
+ </void>
+ <void method="put">
+ <string>transient_lastDdlTime</string>
+ <string>1263595117</string>
+ </void>
+ </object>
+ </void>
+ <void property="serdeClassName">
+ <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ </void>
+ </object>
</void>
</object>
</void>
- <void property="typeInfo">
- <object class="org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo">
- <void property="typeName">
- <string>double</string>
+ <void property="counterNames">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <string>CNTR_NAME_FS_8_NUM_INPUT_ROWS</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_FS_8_NUM_OUTPUT_ROWS</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_FS_8_TIME_TAKEN</string>
+ </void>
+ <void method="add">
+ <string>CNTR_NAME_FS_8_FATAL_ERROR</string>
</void>
</object>
</void>
- </object>
- </void>
- </object>
- </void>
- <void property="tag">
- <int>-1</int>
- </void>
- <void property="valueCols">
- <object class="java.util.ArrayList">
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc">
- <void property="column">
- <string>key</string>
- </void>
- <void property="tabAlias">
- <string></string>
+ <void property="operatorId">
+ <string>FS_8</string>
</void>
- <void property="typeInfo">
- <object id="PrimitiveTypeInfo0" class="org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo">
- <void property="typeName">
- <string>string</string>
+ <void property="parentOperators">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object idref="ExtractOperator0"/>
</void>
</object>
</void>
+ <void property="schema">
+ <object idref="RowSchema0"/>
+ </void>
</object>
</void>
- <void method="add">
+ </object>
+ </void>
+ <void property="conf">
+ <object class="org.apache.hadoop.hive.ql.plan.extractDesc">
+ <void property="col">
<object class="org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc">
<void property="column">
- <string>value</string>
+ <string>VALUE</string>
</void>
<void property="tabAlias">
<string></string>
@@ -267,84 +510,48 @@
</void>
</object>
</void>
- <void property="valueSerializeInfo">
- <object id="tableDesc2" class="org.apache.hadoop.hive.ql.plan.tableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe</class>
+ <void property="counterNames">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <string>CNTR_NAME_OP_7_NUM_INPUT_ROWS</string>
</void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.SequenceFileInputFormat</class>
+ <void method="add">
+ <string>CNTR_NAME_OP_7_NUM_OUTPUT_ROWS</string>
</void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat</class>
+ <void method="add">
+ <string>CNTR_NAME_OP_7_TIME_TAKEN</string>
</void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>columns</string>
- <string>_col0,_col1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string,string</string>
- </void>
- <void method="put">
- <string>escape.delim</string>
- <string>\</string>
- </void>
- </object>
+ <void method="add">
+ <string>CNTR_NAME_OP_7_FATAL_ERROR</string>
</void>
</object>
</void>
- </object>
- </void>
- <void property="counterNames">
- <object class="java.util.ArrayList">
- <void method="add">
- <string>CNTR_NAME_RS_6_NUM_INPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_RS_6_NUM_OUTPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_RS_6_TIME_TAKEN</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_RS_6_FATAL_ERROR</string>
+ <void property="operatorId">
+ <string>OP_7</string>
</void>
- </object>
- </void>
- <void property="operatorId">
- <string>RS_6</string>
- </void>
- <void property="parentOperators">
- <object class="java.util.ArrayList">
- <void method="add">
- <object idref="TableScanOperator0"/>
- </void>
- </object>
- </void>
- <void property="schema">
- <object id="RowSchema0" class="org.apache.hadoop.hive.ql.exec.RowSchema">
- <void property="signature">
- <object class="java.util.Vector">
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
- <void property="internalName">
- <string>key</string>
- </void>
- <void property="type">
- <object idref="PrimitiveTypeInfo0"/>
- </void>
- </object>
- </void>
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
- <void property="internalName">
- <string>value</string>
+ <void property="schema">
+ <object class="org.apache.hadoop.hive.ql.exec.RowSchema">
+ <void property="signature">
+ <object class="java.util.Vector">
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
+ <void property="internalName">
+ <string>0</string>
+ </void>
+ <void property="type">
+ <object idref="PrimitiveTypeInfo0"/>
+ </void>
+ </object>
</void>
- <void property="type">
- <object idref="PrimitiveTypeInfo0"/>
+ <void method="add">
+ <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
+ <void property="internalName">
+ <string>1</string>
+ </void>
+ <void property="type">
+ <object idref="PrimitiveTypeInfo0"/>
+ </void>
+ </object>
</void>
</object>
</void>
@@ -352,250 +559,37 @@
</void>
</object>
</void>
- </object>
- </void>
- </object>
- </void>
- <void property="counterNames">
- <object class="java.util.ArrayList">
- <void method="add">
- <string>CNTR_NAME_TS_5_NUM_INPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_TS_5_NUM_OUTPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_TS_5_TIME_TAKEN</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_TS_5_FATAL_ERROR</string>
- </void>
- </object>
- </void>
- <void property="operatorId">
- <string>TS_5</string>
- </void>
- <void property="schema">
- <object idref="RowSchema0"/>
- </void>
- </object>
- </void>
- </object>
- </void>
- <void property="keyDesc">
- <object idref="tableDesc1"/>
- </void>
- <void property="numReduceTasks">
- <int>-1</int>
- </void>
- <void property="pathToAliases">
- <object class="java.util.LinkedHashMap">
- <void method="put">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
- <object class="java.util.ArrayList">
- <void method="add">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
- </void>
- </object>
- </void>
- </object>
- </void>
- <void property="pathToPartitionInfo">
- <object class="java.util.LinkedHashMap">
- <void method="put">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
- <object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object idref="Properties0"/>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void property="tableDesc">
- <object idref="tableDesc0"/>
- </void>
- </object>
- </void>
- </object>
- </void>
- <void property="reducer">
- <object id="ExtractOperator0" class="org.apache.hadoop.hive.ql.exec.ExtractOperator">
- <void property="childOperators">
- <object class="java.util.ArrayList">
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.FileSinkOperator">
- <void property="conf">
- <object class="org.apache.hadoop.hive.ql.plan.fileSinkDesc">
- <void property="dirName">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1704458189/10000</string>
- </void>
- <void property="tableInfo">
- <object class="org.apache.hadoop.hive.ql.plan.tableDesc">
- <void property="deserializerClass">
- <class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>
- </void>
- <void property="inputFileFormatClass">
- <class>org.apache.hadoop.mapred.TextInputFormat</class>
- </void>
- <void property="outputFileFormatClass">
- <class>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</class>
- </void>
- <void property="properties">
- <object class="java.util.Properties">
- <void method="put">
- <string>name</string>
- <string>dest1</string>
- </void>
- <void method="put">
- <string>columns.types</string>
- <string>string:string</string>
- </void>
- <void method="put">
- <string>serialization.ddl</string>
- <string>struct dest1 { string key, string value}</string>
- </void>
- <void method="put">
- <string>serialization.format</string>
- <string>1</string>
- </void>
- <void method="put">
- <string>columns</string>
- <string>key,value</string>
- </void>
- <void method="put">
- <string>bucket_count</string>
- <string>-1</string>
- </void>
- <void method="put">
- <string>serialization.lib</string>
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
- </void>
- <void method="put">
- <string>file.inputformat</string>
- <string>org.apache.hadoop.mapred.TextInputFormat</string>
- </void>
- <void method="put">
- <string>location</string>
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/dest1</string>
- </void>
- <void method="put">
- <string>file.outputformat</string>
- <string>org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat</string>
- </void>
- <void method="put">
- <string>transient_lastDdlTime</string>
- <string>1263415292</string>
- </void>
- </object>
- </void>
- <void property="serdeClassName">
- <string>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</string>
+ <void property="tagToValueDesc">
+ <object class="java.util.ArrayList">
+ <void method="add">
+ <object idref="tableDesc1"/>
</void>
</object>
</void>
</object>
</void>
- <void property="counterNames">
- <object class="java.util.ArrayList">
- <void method="add">
- <string>CNTR_NAME_FS_8_NUM_INPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_FS_8_NUM_OUTPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_FS_8_TIME_TAKEN</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_FS_8_FATAL_ERROR</string>
- </void>
- </object>
- </void>
- <void property="operatorId">
- <string>FS_8</string>
- </void>
- <void property="parentOperators">
- <object class="java.util.ArrayList">
- <void method="add">
- <object idref="ExtractOperator0"/>
- </void>
- </object>
- </void>
- <void property="schema">
- <object idref="RowSchema0"/>
- </void>
</object>
</void>
</object>
</void>
- <void property="conf">
- <object class="org.apache.hadoop.hive.ql.plan.extractDesc">
- <void property="col">
- <object class="org.apache.hadoop.hive.ql.plan.exprNodeColumnDesc">
- <void property="column">
- <string>VALUE</string>
+ <void property="work">
+ <object class="org.apache.hadoop.hive.ql.plan.moveWork">
+ <void property="loadTableWork">
+ <object class="org.apache.hadoop.hive.ql.plan.loadTableDesc">
+ <void property="partitionSpec">
+ <object class="java.util.HashMap"/>
+ </void>
+ <void property="replace">
+ <boolean>true</boolean>
</void>
- <void property="tabAlias">
- <string></string>
+ <void property="sourceDir">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/200354901/10000</string>
</void>
- <void property="typeInfo">
- <object idref="PrimitiveTypeInfo0"/>
+ <void property="table">
+ <object idref="tableDesc2"/>
</void>
- </object>
- </void>
- </object>
- </void>
- <void property="counterNames">
- <object class="java.util.ArrayList">
- <void method="add">
- <string>CNTR_NAME_OP_7_NUM_INPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_OP_7_NUM_OUTPUT_ROWS</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_OP_7_TIME_TAKEN</string>
- </void>
- <void method="add">
- <string>CNTR_NAME_OP_7_FATAL_ERROR</string>
- </void>
- </object>
- </void>
- <void property="operatorId">
- <string>OP_7</string>
- </void>
- <void property="schema">
- <object class="org.apache.hadoop.hive.ql.exec.RowSchema">
- <void property="signature">
- <object class="java.util.Vector">
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
- <void property="internalName">
- <string>0</string>
- </void>
- <void property="type">
- <object idref="PrimitiveTypeInfo0"/>
- </void>
- </object>
- </void>
- <void method="add">
- <object class="org.apache.hadoop.hive.ql.exec.ColumnInfo">
- <void property="internalName">
- <string>1</string>
- </void>
- <void property="type">
- <object idref="PrimitiveTypeInfo0"/>
- </void>
- </object>
+ <void property="tmpDir">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/200354901/10001</string>
</void>
</object>
</void>
@@ -603,10 +597,23 @@
</void>
</object>
</void>
- <void property="tagToValueDesc">
- <object class="java.util.ArrayList">
- <void method="add">
- <object idref="tableDesc2"/>
+ </object>
+ </void>
+ <void property="id">
+ <string>Stage-4</string>
+ </void>
+ <void property="work">
+ <object id="moveWork0" class="org.apache.hadoop.hive.ql.plan.moveWork">
+ <void property="loadFileWork">
+ <object class="org.apache.hadoop.hive.ql.plan.loadFileDesc">
+ <void property="isDfsDir">
+ <boolean>true</boolean>
+ </void>
+ <void property="sourceDir">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
+ </void>
+ <void property="targetDir">
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/200354901/10000</string>
</void>
</object>
</void>
@@ -614,6 +621,9 @@
</void>
</object>
</void>
+ <void method="add">
+ <object idref="MapRedTask1"/>
+ </void>
</object>
</void>
<void property="parentTasks">
@@ -629,7 +639,7 @@
<void property="resolverCtx">
<object class="org.apache.hadoop.hive.ql.plan.ConditionalResolverMergeFiles$ConditionalResolverMergeFilesCtx">
<void property="dir">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
</void>
<void property="listTasks">
<object idref="ArrayList0"/>
@@ -715,11 +725,11 @@
</void>
<void method="put">
<string>location</string>
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src1</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/test/data/warehouse/src1</string>
</void>
<void method="put">
<string>transient_lastDdlTime</string>
- <string>1263415292</string>
+ <string>1263595117</string>
</void>
</object>
</void>
@@ -775,10 +785,10 @@
<int>1</int>
</void>
<void property="dirName">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/tmp/1232662310/10002</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/tmp/1246984545/10002</string>
</void>
<void property="tableInfo">
- <object idref="tableDesc0"/>
+ <object idref="tableDesc2"/>
</void>
</object>
</void>
@@ -1140,7 +1150,7 @@
<void property="pathToAliases">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src1</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/test/data/warehouse/src1</string>
<object class="java.util.ArrayList">
<void method="add">
<string>src1</string>
@@ -1152,7 +1162,7 @@
<void property="pathToPartitionInfo">
<object class="java.util.LinkedHashMap">
<void method="put">
- <string>file:/data/users/pyang/mstore/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/src1</string>
+ <string>file:/data/users/heyongqiang/hive-trunk/.ptest_3/build/ql/test/data/warehouse/src1</string>
<object class="org.apache.hadoop.hive.ql.plan.partitionDesc">
<void property="deserializerClass">
<class>org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe</class>