You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by Apache Hudson Server <hu...@hudson.apache.org> on 2010/12/19 17:03:12 UTC

Build failed in Hudson: Hive-trunk-h0.20 #447

See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/447/>

------------------------------------------
[...truncated 5845 lines...]
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.hadoop.hive.serde2.thrift.TBinarySortableProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =xfex7fxffxffx15xfex7fxffxffxfdxfex99x96x8dx8cx8bxacx8bx8dx96x91x98xffxfex8cx9ax9cx90x91x9bxacx8bx8dx96x91x98xffxfex7fxffxffxfdxfex99x96x8dx8cx8bxb4x9ax86xffxfex7fxffxffxfexfex8cx9ax9cx90x91x9bxb4x9ax86xffxfex7fxffxffxfdxfex80x00x00xe9xfex40x0fxffxffxffxffxffxffxfexc0x04x00x00x00x00x00x00
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.thrift.protocol.TBinaryProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =x08xffxffx00x00x00xeax0fxffxfex0bx00x00x00x02x00x00x00x0bx66x69x72x73x74x53x74x72x69x6ex67x00x00x00x0cx73x65x63x6fx6ex64x53x74x72x69x6ex67x0dxffxfdx0bx08x00x00x00x02x00x00x00x08x66x69x72x73x74x4bx65x79x00x00x00x01x00x00x00x09x73x65x63x6fx6ex64x4bx65x79x00x00x00x02x08xffxfcxffxffxffx16x04xffxfbx3fxf0x00x00x00x00x00x00x04xffxfaxc0x04x00x00x00x00x00x00x00
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.thrift.protocol.TJSONProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =x7bx22x2dx31x22x3ax7bx22x69x33x32x22x3ax32x33x34x7dx2cx22x2dx32x22x3ax7bx22x6cx73x74x22x3ax5bx22x73x74x72x22x2cx32x2cx22x66x69x72x73x74x53x74x72x69x6ex67x22x2cx22x73x65x63x6fx6ex64x53x74x72x69x6ex67x22x5dx7dx2cx22x2dx33x22x3ax7bx22x6dx61x70x22x3ax5bx22x73x74x72x22x2cx22x69x33x32x22x2cx32x2cx7bx22x66x69x72x73x74x4bx65x79x22x3ax31x2cx22x73x65x63x6fx6ex64x4bx65x79x22x3ax32x7dx5dx7dx2cx22x2dx34x22x3ax7bx22x69x33x32x22x3ax2dx32x33x34x7dx2cx22x2dx35x22x3ax7bx22x64x62x6cx22x3ax31x2ex30x7dx2cx22x2dx36x22x3ax7bx22x64x62x6cx22x3ax2dx32x2ex35x7dx7d
    [junit] bytes in text ={"-1":{"i32":234},"-2":{"lst":["str",2,"firstString","secondString"]},"-3":{"map":["str","i32",2,{"firstKey":1,"secondKey":2}]},"-4":{"i32":-234},"-5":{"dbl":1.0},"-6":{"dbl":-2.5}}
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =x32x33x34x01x66x69x72x73x74x53x74x72x69x6ex67x02x73x65x63x6fx6ex64x53x74x72x69x6ex67x01x66x69x72x73x74x4bx65x79x03x31x02x73x65x63x6fx6ex64x4bx65x79x03x32x01x2dx32x33x34x01x31x2ex30x01x2dx32x2ex35
    [junit] bytes in text =234firstStringsecondStringfirstKey1secondKey2-2341.0-2.5
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Beginning Test testTBinarySortableProtocol:
    [junit] Testing struct test { double hello}
    [junit] Testing struct test { i32 hello}
    [junit] Testing struct test { i64 hello}
    [junit] Testing struct test { string hello}
    [junit] Testing struct test { string hello, double another}
    [junit] Test testTBinarySortableProtocol passed!
    [junit] bytes in text =234	firstStringsecondString	firstKey1secondKey2>
    [junit] compare to    =234	firstStringsecondString	firstKey1secondKey2>
    [junit] o class = class java.util.ArrayList
    [junit] o size = 3
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}]
    [junit] bytes in text =234	firstStringsecondString	firstKey1secondKey2>
    [junit] compare to    =234	firstStringsecondString	firstKey1secondKey2>
    [junit] o class = class java.util.ArrayList
    [junit] o size = 3
    [junit] o = [234, null, {firstKey=1, secondKey=2}]
    [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 0.398 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazy.TestLazyArrayMapStruct
    [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.228 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazy.TestLazyPrimitive
    [junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.194 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazy.TestLazySimpleSerDe
    [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.246 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazybinary.TestLazyBinarySerDe
    [junit] Beginning Test TestLazyBinarySerDe:
    [junit] Test TestLazyBinarySerDe passed!
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.681 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestObjectInspectorConverters
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.217 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestObjectInspectorUtils
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.218 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestProtocolBuffersObjectInspectors
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.201 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestReflectionObjectInspectors
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.202 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestStandardObjectInspectors
    [junit] [null, null, null]
    [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 0.232 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestThriftObjectInspectors
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.214 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestUnionStructObjectInspector
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.2 sec

test-conditions:

gen-test:

create-dirs:

compile-ant-tasks:

create-dirs:

init:

compile:
     [echo] Compiling: anttasks
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

deploy-ant-tasks:

create-dirs:

init:

compile:
     [echo] Compiling: anttasks
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

jar:

init:

metastore-init:

build-grammar:

model-compile:
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/metastore/build.xml>:69: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

ivy-init-dirs:

ivy-download:
      [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
      [get] To: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
      [get] Not modified - so not downloaded

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-resolve:
[ivy:resolve] :: Ivy 2.1.0 - 20090925235825 :: http://ant.apache.org/ivy/ ::
[ivy:resolve] :: loading settings :: file = <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ivy/ivysettings.xml>

ivy-retrieve:

core-compile:
     [echo] Compiling: 
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/metastore/build.xml>:60: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

model-enhance:
[datanucleusenhancer] log4j:WARN No appenders could be found for logger (DataNucleus.Enhancer).
[datanucleusenhancer] log4j:WARN Please initialize the log4j system properly.
[datanucleusenhancer] DataNucleus Enhancer (version 2.0.3) : Enhancement of classes
[datanucleusenhancer] DataNucleus Enhancer : Classpath
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/common/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/serde/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/cli/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/shims/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hwi/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hadoopcore/hadoop-0.20.0/hadoop-0.20.0-core.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hadoopcore/hadoop-0.20.0/hadoop-0.20.0-tools.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/anttasks/hive-anttasks-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/cli/hive-cli-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/common/hive-common-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/contrib/hive_contrib.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hbase-handler/hive_hbase-handler.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hwi/hive-hwi-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/jdbc/hive-jdbc-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/hive-metastore-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/hive-exec-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/serde/hive-serde-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/hive-service-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/shims/hive-shims-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/asm-3.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-cli-2.0-SNAPSHOT.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-codec-1.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-collections-3.2.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-lang-2.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-logging-1.0.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-logging-api-1.0.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/derby.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/hbase-0.20.3-test.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/hbase-0.20.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/json.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/log4j-1.2.15.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/thrift-0.5.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/thrift-fb303-0.5.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/velocity-1.5.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/zookeeper-3.2.2.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/antlr-2.7.7.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/antlr-3.0.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/antlr-runtime-3.0.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/stringtemplate-3.1b1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/commons-dbcp-1.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/commons-pool-1.5.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-connectionpool-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-core-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-enhancer-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-rdbms-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/jdo2-api-2.3-ec.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/log4j-1.2.16.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/slf4j-api-1.6.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/slf4j-log4j12-1.6.1.jar>
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MDatabase
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MFieldSchema
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MType
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MTable
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MSerDeInfo
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MOrder
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MStorageDescriptor
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MPartition
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MIndex
[datanucleusenhancer] DataNucleus Enhancer completed with success for 9 classes. Timings : input=315 ms, enhance=72 ms, total=387 ms. Consult the log for full details

compile:

compile-test:
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:317: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds
    [javac] Compiling 2 source files to <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/classes>
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:330: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

test-jar:
      [jar] Building MANIFEST-only jar: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/test-udfs.jar>

test-init:
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/data>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/clientpositive>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/clientnegative>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/positive>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/negative>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/data/warehouse>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/data/metadb>

test:
    [junit] Running org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore
    [junit] Tests run: 11, Failures: 0, Errors: 1, Time elapsed: 8.96 sec
    [junit] Test org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore FAILED

BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:209: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:132: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:454: Tests failed!

Total time: 1 minute 56 seconds
Recording test results


Hudson build is back to normal : Hive-trunk-h0.20 #453

Posted by Apache Hudson Server <hu...@hudson.apache.org>.
See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/453/changes>



Build failed in Hudson: Hive-trunk-h0.20 #452

Posted by Apache Hudson Server <hu...@hudson.apache.org>.
See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/452/changes>

Changes:

[namit] HIVE-1806 Merge per dynamic partition based on size of each dynamic partition
(Ning Zhang via namit)

[namit] HIVE-1456 No need to check for LOG as null in sort-merge join
(Alexey Diomin via namit)

------------------------------------------
[...truncated 18843 lines...]
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table1.q.out>
    [junit] Done query: unknown_table1.q
    [junit] Begin query: unknown_table2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table2.q.out>
    [junit] Done query: unknown_table2.q
    [junit] Begin query: wrong_distinct1.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct1.q.out>
    [junit] Done query: wrong_distinct1.q
    [junit] Begin query: wrong_distinct2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct2.q.out>
    [junit] Done query: wrong_distinct2.q
    [junit] Cleaning up TestParseNegative
    [junit] Tests run: 32, Failures: 0, Errors: 0, Time elapsed: 254.645 sec
    [junit] Running org.apache.hadoop.hive.ql.tool.TestLineageInfo
    [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 0.419 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateAdd
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.193 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateDiff
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.178 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateSub
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.158 sec
    [junit] Running org.apache.hadoop.hive.ql.util.TestDosToUnix
    [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 0.056 sec

BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:209: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:132: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:454: Tests failed!

Total time: 207 minutes 52 seconds
Recording test results


Build failed in Hudson: Hive-trunk-h0.20 #451

Posted by Apache Hudson Server <hu...@hudson.apache.org>.
See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/451/changes>

Changes:

[jssarma] HIVE-1852 Reduce unnecessary DFSClient.rename() calls (Ning Zhang via jssarma)

------------------------------------------
[...truncated 14785 lines...]
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table1.q.out>
    [junit] Done query: unknown_table1.q
    [junit] Begin query: unknown_table2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table2.q.out>
    [junit] Done query: unknown_table2.q
    [junit] Begin query: wrong_distinct1.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct1.q.out>
    [junit] Done query: wrong_distinct1.q
    [junit] Begin query: wrong_distinct2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct2.q.out>
    [junit] Done query: wrong_distinct2.q
    [junit] Cleaning up TestParseNegative
    [junit] Tests run: 32, Failures: 0, Errors: 0, Time elapsed: 221.054 sec
    [junit] Running org.apache.hadoop.hive.ql.tool.TestLineageInfo
    [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 0.401 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateAdd
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.151 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateDiff
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.151 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateSub
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.152 sec
    [junit] Running org.apache.hadoop.hive.ql.util.TestDosToUnix
    [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 0.05 sec

BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:209: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:132: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:454: Tests failed!

Total time: 179 minutes 35 seconds
Recording test results


Build failed in Hudson: Hive-trunk-h0.20 #450

Posted by Apache Hudson Server <hu...@hudson.apache.org>.
See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/450/changes>

Changes:

[jvs] HIVE-1856 Implement DROP TABLE/VIEW IF EXISTS
(Marcel Kornacker via jvs)

------------------------------------------
[...truncated 15125 lines...]
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table1.q.out>
    [junit] Done query: unknown_table1.q
    [junit] Begin query: unknown_table2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table2.q.out>
    [junit] Done query: unknown_table2.q
    [junit] Begin query: wrong_distinct1.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct1.q.out>
    [junit] Done query: wrong_distinct1.q
    [junit] Begin query: wrong_distinct2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct2.q.out>
    [junit] Done query: wrong_distinct2.q
    [junit] Cleaning up TestParseNegative
    [junit] Tests run: 32, Failures: 0, Errors: 0, Time elapsed: 227.719 sec
    [junit] Running org.apache.hadoop.hive.ql.tool.TestLineageInfo
    [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 0.362 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateAdd
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.147 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateDiff
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.148 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateSub
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.149 sec
    [junit] Running org.apache.hadoop.hive.ql.util.TestDosToUnix
    [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 0.05 sec

BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:209: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:132: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:454: Tests failed!

Total time: 177 minutes 40 seconds
Recording test results


Build failed in Hudson: Hive-trunk-h0.20 #449

Posted by Apache Hudson Server <hu...@hudson.apache.org>.
See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/449/changes>

Changes:

[pauly] HIVE-1857 mixed case tablename on lefthand side of LATERAL VIEW results in
query failing with confusing error message (John Sichi via pauly)

[namit] HIVE-1855 Include Process ID in the log4j log file name
(Ning Zhang via namit)

[nzhang] HIVE-1835. Better auto-complete for Hive (Paul Butler via Ning Zhang)

[jssarma] commit second diff for hive-1846 (rvadali via jssarma)

[jssarma] Reversing erroneous commit

[jssarma] commit second diff for hive-1846 (rvadali via jssarma)

[namit] HIVE-1854 Temporarily disable metastore tests for listPartitionsByFilter()
(Paul Yang via namit)

------------------------------------------
[...truncated 15104 lines...]
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table1.q.out>
    [junit] Done query: unknown_table1.q
    [junit] Begin query: unknown_table2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/unknown_table2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/unknown_table2.q.out>
    [junit] Done query: unknown_table2.q
    [junit] Begin query: wrong_distinct1.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct1.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct1.q.out>
    [junit] Done query: wrong_distinct1.q
    [junit] Begin query: wrong_distinct2.q
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-08/hr=12
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=11>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=11
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
    [junit] rmr: cannot remove p<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12>: No such file or directory.
    [junit] POSTHOOK: Output: default@srcpart@ds=2008-04-09/hr=12
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket0.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket1.txt>
    [junit] Loading data to table srcbucket
    [junit] POSTHOOK: Output: default@srcbucket
    [junit] OK
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket20.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket21.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket22.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/srcbucket23.txt>
    [junit] Loading data to table srcbucket2
    [junit] POSTHOOK: Output: default@srcbucket2
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table src
    [junit] POSTHOOK: Output: default@src
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv3.txt>
    [junit] Loading data to table src1
    [junit] POSTHOOK: Output: default@src1
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.seq>
    [junit] Loading data to table src_sequencefile
    [junit] POSTHOOK: Output: default@src_sequencefile
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/complex.seq>
    [junit] Loading data to table src_thrift
    [junit] POSTHOOK: Output: default@src_thrift
    [junit] OK
    [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/json.txt>
    [junit] Loading data to table src_json
    [junit] POSTHOOK: Output: default@src_json
    [junit] OK
    [junit] diff <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/test/logs/negative/wrong_distinct2.q.out> <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/src/test/results/compiler/errors/wrong_distinct2.q.out>
    [junit] Done query: wrong_distinct2.q
    [junit] Cleaning up TestParseNegative
    [junit] Tests run: 32, Failures: 0, Errors: 0, Time elapsed: 222.469 sec
    [junit] Running org.apache.hadoop.hive.ql.tool.TestLineageInfo
    [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 0.343 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateAdd
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.151 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateDiff
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.156 sec
    [junit] Running org.apache.hadoop.hive.ql.udf.TestUDFDateSub
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.152 sec
    [junit] Running org.apache.hadoop.hive.ql.util.TestDosToUnix
    [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 0.051 sec

BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:209: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:132: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:454: Tests failed!

Total time: 172 minutes 10 seconds
Recording test results


Build failed in Hudson: Hive-trunk-h0.20 #448

Posted by Apache Hudson Server <hu...@hudson.apache.org>.
See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/448/>

------------------------------------------
[...truncated 5845 lines...]
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.hadoop.hive.serde2.thrift.TBinarySortableProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =xfex7fxffxffx15xfex7fxffxffxfdxfex99x96x8dx8cx8bxacx8bx8dx96x91x98xffxfex8cx9ax9cx90x91x9bxacx8bx8dx96x91x98xffxfex7fxffxffxfdxfex99x96x8dx8cx8bxb4x9ax86xffxfex7fxffxffxfexfex8cx9ax9cx90x91x9bxb4x9ax86xffxfex7fxffxffxfdxfex80x00x00xe9xfex40x0fxffxffxffxffxffxffxfexc0x04x00x00x00x00x00x00
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.thrift.protocol.TBinaryProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =x08xffxffx00x00x00xeax0fxffxfex0bx00x00x00x02x00x00x00x0bx66x69x72x73x74x53x74x72x69x6ex67x00x00x00x0cx73x65x63x6fx6ex64x53x74x72x69x6ex67x0dxffxfdx0bx08x00x00x00x02x00x00x00x08x66x69x72x73x74x4bx65x79x00x00x00x01x00x00x00x09x73x65x63x6fx6ex64x4bx65x79x00x00x00x02x08xffxfcxffxffxffx16x04xffxfbx3fxf0x00x00x00x00x00x00x04xffxfaxc0x04x00x00x00x00x00x00x00
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.thrift.protocol.TJSONProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =x7bx22x2dx31x22x3ax7bx22x69x33x32x22x3ax32x33x34x7dx2cx22x2dx32x22x3ax7bx22x6cx73x74x22x3ax5bx22x73x74x72x22x2cx32x2cx22x66x69x72x73x74x53x74x72x69x6ex67x22x2cx22x73x65x63x6fx6ex64x53x74x72x69x6ex67x22x5dx7dx2cx22x2dx33x22x3ax7bx22x6dx61x70x22x3ax5bx22x73x74x72x22x2cx22x69x33x32x22x2cx32x2cx7bx22x66x69x72x73x74x4bx65x79x22x3ax31x2cx22x73x65x63x6fx6ex64x4bx65x79x22x3ax32x7dx5dx7dx2cx22x2dx34x22x3ax7bx22x69x33x32x22x3ax2dx32x33x34x7dx2cx22x2dx35x22x3ax7bx22x64x62x6cx22x3ax31x2ex30x7dx2cx22x2dx36x22x3ax7bx22x64x62x6cx22x3ax2dx32x2ex35x7dx7d
    [junit] bytes in text ={"-1":{"i32":234},"-2":{"lst":["str",2,"firstString","secondString"]},"-3":{"map":["str","i32",2,{"firstKey":1,"secondKey":2}]},"-4":{"i32":-234},"-5":{"dbl":1.0},"-6":{"dbl":-2.5}}
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Testing protocol: org.apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol
    [junit] TypeName = struct<_hello:int,2bye:array<string>,another:map<string,int>,nhello:int,d:double,nd:double>
    [junit] bytes =x32x33x34x01x66x69x72x73x74x53x74x72x69x6ex67x02x73x65x63x6fx6ex64x53x74x72x69x6ex67x01x66x69x72x73x74x4bx65x79x03x31x02x73x65x63x6fx6ex64x4bx65x79x03x32x01x2dx32x33x34x01x31x2ex30x01x2dx32x2ex35
    [junit] bytes in text =234firstStringsecondStringfirstKey1secondKey2-2341.0-2.5
    [junit] o class = class java.util.ArrayList
    [junit] o size = 6
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}, -234, 1.0, -2.5]
    [junit] Beginning Test testTBinarySortableProtocol:
    [junit] Testing struct test { double hello}
    [junit] Testing struct test { i32 hello}
    [junit] Testing struct test { i64 hello}
    [junit] Testing struct test { string hello}
    [junit] Testing struct test { string hello, double another}
    [junit] Test testTBinarySortableProtocol passed!
    [junit] bytes in text =234	firstStringsecondString	firstKey1secondKey2>
    [junit] compare to    =234	firstStringsecondString	firstKey1secondKey2>
    [junit] o class = class java.util.ArrayList
    [junit] o size = 3
    [junit] o[0] class = class java.lang.Integer
    [junit] o[1] class = class java.util.ArrayList
    [junit] o[2] class = class java.util.HashMap
    [junit] o = [234, [firstString, secondString], {firstKey=1, secondKey=2}]
    [junit] bytes in text =234	firstStringsecondString	firstKey1secondKey2>
    [junit] compare to    =234	firstStringsecondString	firstKey1secondKey2>
    [junit] o class = class java.util.ArrayList
    [junit] o size = 3
    [junit] o = [234, null, {firstKey=1, secondKey=2}]
    [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 0.446 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazy.TestLazyArrayMapStruct
    [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.259 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazy.TestLazyPrimitive
    [junit] Tests run: 8, Failures: 0, Errors: 0, Time elapsed: 0.214 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazy.TestLazySimpleSerDe
    [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.253 sec
    [junit] Running org.apache.hadoop.hive.serde2.lazybinary.TestLazyBinarySerDe
    [junit] Beginning Test TestLazyBinarySerDe:
    [junit] Test TestLazyBinarySerDe passed!
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.725 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestObjectInspectorConverters
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.213 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestObjectInspectorUtils
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.226 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestProtocolBuffersObjectInspectors
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.203 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestReflectionObjectInspectors
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.204 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestStandardObjectInspectors
    [junit] [null, null, null]
    [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 0.263 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestThriftObjectInspectors
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.258 sec
    [junit] Running org.apache.hadoop.hive.serde2.objectinspector.TestUnionStructObjectInspector
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.2 sec

test-conditions:

gen-test:

create-dirs:

compile-ant-tasks:

create-dirs:

init:

compile:
     [echo] Compiling: anttasks
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

deploy-ant-tasks:

create-dirs:

init:

compile:
     [echo] Compiling: anttasks
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

jar:

init:

metastore-init:

build-grammar:

model-compile:
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/metastore/build.xml>:69: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

ivy-init-dirs:

ivy-download:
      [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
      [get] To: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
      [get] Not modified - so not downloaded

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-resolve:
[ivy:resolve] :: Ivy 2.1.0 - 20090925235825 :: http://ant.apache.org/ivy/ ::
[ivy:resolve] :: loading settings :: file = <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ivy/ivysettings.xml>

ivy-retrieve:

core-compile:
     [echo] Compiling: 
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/metastore/build.xml>:60: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

model-enhance:
[datanucleusenhancer] log4j:WARN No appenders could be found for logger (DataNucleus.Enhancer).
[datanucleusenhancer] log4j:WARN Please initialize the log4j system properly.
[datanucleusenhancer] DataNucleus Enhancer (version 2.0.3) : Enhancement of classes
[datanucleusenhancer] DataNucleus Enhancer : Classpath
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/common/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/serde/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/cli/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/shims/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hwi/classes>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hadoopcore/hadoop-0.20.0/hadoop-0.20.0-core.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hadoopcore/hadoop-0.20.0/hadoop-0.20.0-tools.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/anttasks/hive-anttasks-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/cli/hive-cli-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/common/hive-common-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/contrib/hive_contrib.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hbase-handler/hive_hbase-handler.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hwi/hive-hwi-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/jdbc/hive-jdbc-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/hive-metastore-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ql/hive-exec-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/serde/hive-serde-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/hive-service-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/shims/hive-shims-0.7.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/asm-3.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-cli-2.0-SNAPSHOT.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-codec-1.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-collections-3.2.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-lang-2.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-logging-1.0.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/commons-logging-api-1.0.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/derby.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/hbase-0.20.3-test.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/hbase-0.20.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/json.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/log4j-1.2.15.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/thrift-0.5.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/thrift-fb303-0.5.0.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/velocity-1.5.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/lib/zookeeper-3.2.2.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/antlr-2.7.7.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/antlr-3.0.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/antlr-runtime-3.0.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ql/lib/stringtemplate-3.1b1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/commons-dbcp-1.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/commons-pool-1.5.4.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-connectionpool-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-core-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-enhancer-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/datanucleus-rdbms-2.0.3.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/jdo2-api-2.3-ec.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/log4j-1.2.16.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/slf4j-api-1.6.1.jar>
[datanucleusenhancer] >>  <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/default/slf4j-log4j12-1.6.1.jar>
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MDatabase
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MFieldSchema
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MType
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MTable
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MSerDeInfo
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MOrder
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MStorageDescriptor
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MPartition
[datanucleusenhancer] ENHANCED (PersistenceCapable) : org.apache.hadoop.hive.metastore.model.MIndex
[datanucleusenhancer] DataNucleus Enhancer completed with success for 9 classes. Timings : input=373 ms, enhance=83 ms, total=456 ms. Consult the log for full details

compile:

compile-test:
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:317: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds
    [javac] Compiling 2 source files to <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/classes>
    [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:330: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds

test-jar:
      [jar] Building MANIFEST-only jar: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/test-udfs.jar>

test-init:
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/data>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/clientpositive>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/clientnegative>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/positive>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/logs/negative>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/data/warehouse>
    [mkdir] Created dir: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/metastore/test/data/metadb>

test:
    [junit] Running org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore
    [junit] Tests run: 11, Failures: 0, Errors: 1, Time elapsed: 9.382 sec
    [junit] Test org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore FAILED

BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:209: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:132: The following error occurred while executing this line:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:454: Tests failed!

Total time: 1 minute 58 seconds
Recording test results