You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2014/12/10 22:04:46 UTC

svn commit: r1644522 - in /hive/branches/spark: itests/src/test/resources/testconfiguration.properties ql/src/test/results/clientpositive/spark/groupby_bigdata.q.out ql/src/test/results/clientpositive/spark/rcfile_bigdata.q.out

Author: xuefu
Date: Wed Dec 10 21:04:45 2014
New Revision: 1644522

URL: http://svn.apache.org/r1644522
Log:
HIVE-8507: UT: fix rcfile_bigdata test [Spark Branch] (Chinna via Xuefu)

Added:
    hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_bigdata.q.out
    hive/branches/spark/ql/src/test/results/clientpositive/spark/rcfile_bigdata.q.out
Modified:
    hive/branches/spark/itests/src/test/resources/testconfiguration.properties

Modified: hive/branches/spark/itests/src/test/resources/testconfiguration.properties
URL: http://svn.apache.org/viewvc/hive/branches/spark/itests/src/test/resources/testconfiguration.properties?rev=1644522&r1=1644521&r2=1644522&view=diff
==============================================================================
--- hive/branches/spark/itests/src/test/resources/testconfiguration.properties (original)
+++ hive/branches/spark/itests/src/test/resources/testconfiguration.properties Wed Dec 10 21:04:45 2014
@@ -579,6 +579,7 @@ spark.query.files=add_part_multiple.q, \
   groupby9.q, \
   groupby10.q, \
   groupby11.q, \
+  groupby_bigdata.q, \
   groupby_complex_types.q, \
   groupby_complex_types_multi_single_reducer.q, \
   groupby_cube1.q, \
@@ -759,6 +760,7 @@ spark.query.files=add_part_multiple.q, \
   ptf_register_tblfn.q, \
   ptf_seqfile.q, \
   ptf_streaming.q, \
+  rcfile_bigdata.q, \
   reduce_deduplicate_exclude_join.q, \
   router_join_ppr.q, \
   sample1.q, \

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_bigdata.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_bigdata.q.out?rev=1644522&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_bigdata.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_bigdata.q.out Wed Dec 10 21:04:45 2014
@@ -0,0 +1,11 @@
+PREHOOK: query: select count(distinct subq.key) from
+(FROM src MAP src.key USING 'python dumpdata_script.py' AS key WHERE src.key = 10) subq
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+#### A masked pattern was here ####
+POSTHOOK: query: select count(distinct subq.key) from
+(FROM src MAP src.key USING 'python dumpdata_script.py' AS key WHERE src.key = 10) subq
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+#### A masked pattern was here ####
+1000022

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/rcfile_bigdata.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/rcfile_bigdata.q.out?rev=1644522&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/rcfile_bigdata.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/rcfile_bigdata.q.out Wed Dec 10 21:04:45 2014
@@ -0,0 +1,47 @@
+PREHOOK: query: CREATE table columnTable_Bigdata (key STRING, value STRING)
+ROW FORMAT SERDE
+  'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
+STORED AS
+  INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat'
+  OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@columnTable_Bigdata
+POSTHOOK: query: CREATE table columnTable_Bigdata (key STRING, value STRING)
+ROW FORMAT SERDE
+  'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe'
+STORED AS
+  INPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileInputFormat'
+  OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.RCFileOutputFormat'
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@columnTable_Bigdata
+PREHOOK: query: FROM (FROM src MAP src.key,src.value USING 'python dumpdata_script.py' AS (key,value) WHERE src.key = 10) subq
+INSERT OVERWRITE TABLE columnTable_Bigdata SELECT subq.key, subq.value
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@columntable_bigdata
+POSTHOOK: query: FROM (FROM src MAP src.key,src.value USING 'python dumpdata_script.py' AS (key,value) WHERE src.key = 10) subq
+INSERT OVERWRITE TABLE columnTable_Bigdata SELECT subq.key, subq.value
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@columntable_bigdata
+POSTHOOK: Lineage: columntable_bigdata.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: columntable_bigdata.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: describe columnTable_Bigdata
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@columntable_bigdata
+POSTHOOK: query: describe columnTable_Bigdata
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@columntable_bigdata
+key                 	string              	                    
+value               	string              	                    
+PREHOOK: query: select count(columnTable_Bigdata.key) from columnTable_Bigdata
+PREHOOK: type: QUERY
+PREHOOK: Input: default@columntable_bigdata
+#### A masked pattern was here ####
+POSTHOOK: query: select count(columnTable_Bigdata.key) from columnTable_Bigdata
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@columntable_bigdata
+#### A masked pattern was here ####
+5005500