You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2011/08/10 06:36:14 UTC

Build failed in Jenkins: Hive-trunk-h0.21 #885

See <https://builds.apache.org/job/Hive-trunk-h0.21/885/changes>

Changes:

[pauly] HIVE-2246. Dedupe tables' column schemas from partitions in the metastore db (Sohan Jain via pauly)

------------------------------------------
[...truncated 33777 lines...]

mvn-taskdef:

maven-publish-artifact:
[artifact:install-provider] Installing provider: org.apache.maven.wagon:wagon-http:jar:1.0-beta-2:runtime
[artifact:deploy] Deploying to https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] [INFO] Retrieving previous build number from apache.snapshots.https
[artifact:deploy] Uploading: org/apache/hive/hive-metastore/0.8.0-SNAPSHOT/hive-metastore-0.8.0-20110810.043604-64.jar to repository apache.snapshots.https at https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] Transferring 1664K from apache.snapshots.https
[artifact:deploy] Uploaded 1664K
[artifact:deploy] [INFO] Uploading project information for hive-metastore 0.8.0-20110810.043604-64
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'snapshot org.apache.hive:hive-metastore:0.8.0-SNAPSHOT'
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'artifact org.apache.hive:hive-metastore'

ivy-init-dirs:

ivy-download:
      [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
      [get] To: <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
      [get] Not modified - so not downloaded

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-resolve-maven-ant-tasks:
[ivy:resolve] :: loading settings :: file = <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/ivy/ivysettings.xml>

ivy-retrieve-maven-ant-tasks:
[ivy:cachepath] DEPRECATED: 'ivy.conf.file' is deprecated, use 'ivy.settings.file' instead
[ivy:cachepath] :: loading settings :: file = <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/ivy/ivysettings.xml>

mvn-taskdef:

maven-publish-artifact:
[artifact:install-provider] Installing provider: org.apache.maven.wagon:wagon-http:jar:1.0-beta-2:runtime
[artifact:deploy] Deploying to https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] [INFO] Retrieving previous build number from apache.snapshots.https
[artifact:deploy] Uploading: org/apache/hive/hive-serde/0.8.0-SNAPSHOT/hive-serde-0.8.0-20110810.043606-64.jar to repository apache.snapshots.https at https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] Transferring 453K from apache.snapshots.https
[artifact:deploy] Uploaded 453K
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'artifact org.apache.hive:hive-serde'
[artifact:deploy] [INFO] Uploading project information for hive-serde 0.8.0-20110810.043606-64
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'snapshot org.apache.hive:hive-serde:0.8.0-SNAPSHOT'

ivy-init-dirs:

ivy-download:
      [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
      [get] To: <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
      [get] Not modified - so not downloaded

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-resolve-maven-ant-tasks:
[ivy:resolve] :: loading settings :: file = <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/ivy/ivysettings.xml>

ivy-retrieve-maven-ant-tasks:
[ivy:cachepath] DEPRECATED: 'ivy.conf.file' is deprecated, use 'ivy.settings.file' instead
[ivy:cachepath] :: loading settings :: file = <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/ivy/ivysettings.xml>

mvn-taskdef:

maven-publish-artifact:
[artifact:install-provider] Installing provider: org.apache.maven.wagon:wagon-http:jar:1.0-beta-2:runtime
[artifact:deploy] Deploying to https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] [INFO] Retrieving previous build number from apache.snapshots.https
[artifact:deploy] Uploading: org/apache/hive/hive-service/0.8.0-SNAPSHOT/hive-service-0.8.0-20110810.043609-64.jar to repository apache.snapshots.https at https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] Transferring 170K from apache.snapshots.https
[artifact:deploy] Uploaded 170K
[artifact:deploy] [INFO] Uploading project information for hive-service 0.8.0-20110810.043609-64
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'artifact org.apache.hive:hive-service'
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'snapshot org.apache.hive:hive-service:0.8.0-SNAPSHOT'

ivy-init-dirs:

ivy-download:
      [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
      [get] To: <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
      [get] Not modified - so not downloaded

ivy-probe-antlib:

ivy-init-antlib:

ivy-init:

ivy-resolve-maven-ant-tasks:
[ivy:resolve] :: loading settings :: file = <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/ivy/ivysettings.xml>

ivy-retrieve-maven-ant-tasks:
[ivy:cachepath] DEPRECATED: 'ivy.conf.file' is deprecated, use 'ivy.settings.file' instead
[ivy:cachepath] :: loading settings :: file = <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/ivy/ivysettings.xml>

mvn-taskdef:

maven-publish-artifact:
[artifact:install-provider] Installing provider: org.apache.maven.wagon:wagon-http:jar:1.0-beta-2:runtime
[artifact:deploy] Deploying to https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] [INFO] Retrieving previous build number from apache.snapshots.https
[artifact:deploy] Uploading: org/apache/hive/hive-shims/0.8.0-SNAPSHOT/hive-shims-0.8.0-20110810.043611-64.jar to repository apache.snapshots.https at https://repository.apache.org/content/repositories/snapshots
[artifact:deploy] Transferring 76K from apache.snapshots.https
[artifact:deploy] Uploaded 76K
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'snapshot org.apache.hive:hive-shims:0.8.0-SNAPSHOT'
[artifact:deploy] [INFO] Retrieving previous metadata from apache.snapshots.https
[artifact:deploy] [INFO] Uploading repository metadata for: 'artifact org.apache.hive:hive-shims'
[artifact:deploy] [INFO] Uploading project information for hive-shims 0.8.0-20110810.043611-64

BUILD SUCCESSFUL
Total time: 244 minutes 11 seconds
Archiving artifacts
ERROR: Failed to archive artifacts: hive/build/hive*.tar.gz*
hudson.util.IOException2: java.lang.NoClassDefFoundError: Could not initialize class org.apache.tools.ant.PropertyHelper
	at hudson.FilePath.copyRecursiveTo(FilePath.java:1581)
	at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:117)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
	at hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:662)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:638)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:616)
	at hudson.model.Build$RunnerImpl.post2(Build.java:161)
	at hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:585)
	at hudson.model.Run.run(Run.java:1398)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:145)
Caused by: java.util.concurrent.ExecutionException: java.lang.NoClassDefFoundError: Could not initialize class org.apache.tools.ant.PropertyHelper
	at hudson.remoting.Channel$2.adapt(Channel.java:694)
	at hudson.remoting.Channel$2.adapt(Channel.java:689)
	at hudson.remoting.FutureAdapter.get(FutureAdapter.java:55)
	at hudson.FilePath.copyRecursiveTo(FilePath.java:1579)
	... 11 more
Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.apache.tools.ant.PropertyHelper
	at org.apache.tools.ant.Project.getReference(Project.java:1989)
	at org.apache.tools.ant.ComponentHelper.getComponentHelper(ComponentHelper.java:150)
	at org.apache.tools.ant.ComponentHelper.getElementName(ComponentHelper.java:646)
	at org.apache.tools.ant.types.DataType.getDataTypeName(DataType.java:95)
	at org.apache.tools.ant.types.AbstractFileSet.setupDirectoryScanner(AbstractFileSet.java:520)
	at org.apache.tools.ant.types.AbstractFileSet.getDirectoryScanner(AbstractFileSet.java:483)
	at hudson.util.DirScanner$Glob.scan(DirScanner.java:101)
	at hudson.FilePath.writeToTar(FilePath.java:1617)
	at hudson.FilePath.access$900(FilePath.java:164)
	at hudson.FilePath$33.invoke(FilePath.java:1558)
	at hudson.FilePath$33.invoke(FilePath.java:1555)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1979)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:270)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
	at java.util.concurrent.FutureTask.run(FutureTask.java:138)
	at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
	at java.lang.Thread.run(Thread.java:619)
Recording test results
ERROR: Failed to archive test reports
hudson.util.IOException2: remote file operation failed: <https://builds.apache.org/job/Hive-trunk-h0.21/ws/> at hudson.remoting.Channel@a9aae50:ubuntu1
	at hudson.FilePath.act(FilePath.java:754)
	at hudson.FilePath.act(FilePath.java:740)
	at hudson.tasks.junit.JUnitParser.parse(JUnitParser.java:83)
	at hudson.tasks.junit.JUnitResultArchiver.parse(JUnitResultArchiver.java:123)
	at hudson.tasks.junit.JUnitResultArchiver.perform(JUnitResultArchiver.java:135)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
	at hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:662)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:638)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:616)
	at hudson.model.Build$RunnerImpl.post2(Build.java:161)
	at hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:585)
	at hudson.model.Run.run(Run.java:1398)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:145)
Caused by: java.io.IOException: Remote call on ubuntu1 failed
	at hudson.remoting.Channel.call(Channel.java:670)
	at hudson.FilePath.act(FilePath.java:747)
	... 14 more
Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.apache.tools.ant.PropertyHelper
	at org.apache.tools.ant.Project.getReference(Project.java:1989)
	at org.apache.tools.ant.ComponentHelper.getComponentHelper(ComponentHelper.java:150)
	at org.apache.tools.ant.ComponentHelper.getElementName(ComponentHelper.java:646)
	at org.apache.tools.ant.types.DataType.getDataTypeName(DataType.java:95)
	at org.apache.tools.ant.types.AbstractFileSet.setupDirectoryScanner(AbstractFileSet.java:520)
	at org.apache.tools.ant.types.AbstractFileSet.getDirectoryScanner(AbstractFileSet.java:483)
	at org.apache.tools.ant.types.AbstractFileSet.getDirectoryScanner(AbstractFileSet.java:451)
	at hudson.tasks.junit.JUnitParser$ParseResultCallable.invoke(JUnitParser.java:105)
	at hudson.tasks.junit.JUnitParser$ParseResultCallable.invoke(JUnitParser.java:87)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1979)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:270)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
	at java.util.concurrent.FutureTask.run(FutureTask.java:138)
	at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
	at java.lang.Thread.run(Thread.java:619)
ERROR: Publisher hudson.plugins.violations.ViolationsPublisher aborted due to exception
hudson.util.IOException2: remote file operation failed: <https://builds.apache.org/job/Hive-trunk-h0.21/ws/> at hudson.remoting.Channel@a9aae50:ubuntu1
	at hudson.FilePath.act(FilePath.java:754)
	at hudson.FilePath.act(FilePath.java:740)
	at hudson.plugins.violations.ViolationsPublisher.perform(ViolationsPublisher.java:74)
	at hudson.tasks.BuildStepMonitor$3.perform(BuildStepMonitor.java:36)
	at hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:662)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:638)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:616)
	at hudson.model.Build$RunnerImpl.post2(Build.java:161)
	at hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:585)
	at hudson.model.Run.run(Run.java:1398)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:145)
Caused by: java.io.IOException: Remote call on ubuntu1 failed
	at hudson.remoting.Channel.call(Channel.java:670)
	at hudson.FilePath.act(FilePath.java:747)
	... 12 more
Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.apache.tools.ant.PropertyHelper
	at org.apache.tools.ant.Project.getReference(Project.java:1989)
	at org.apache.tools.ant.ComponentHelper.getComponentHelper(ComponentHelper.java:150)
	at org.apache.tools.ant.ComponentHelper.getElementName(ComponentHelper.java:646)
	at org.apache.tools.ant.types.DataType.getDataTypeName(DataType.java:95)
	at org.apache.tools.ant.types.AbstractFileSet.setupDirectoryScanner(AbstractFileSet.java:520)
	at org.apache.tools.ant.types.AbstractFileSet.getDirectoryScanner(AbstractFileSet.java:483)
	at hudson.plugins.violations.ViolationsCollector.findFiles(ViolationsCollector.java:213)
	at hudson.plugins.violations.ViolationsCollector.doType(ViolationsCollector.java:173)
	at hudson.plugins.violations.ViolationsCollector.invoke(ViolationsCollector.java:114)
	at hudson.plugins.violations.ViolationsCollector.invoke(ViolationsCollector.java:25)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1979)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:270)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
	at java.util.concurrent.FutureTask.run(FutureTask.java:138)
	at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
	at java.lang.Thread.run(Thread.java:619)
Updating HIVE-2246


Jenkins build is back to normal : Hive-trunk-h0.21 #887

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Hive-trunk-h0.21/887/changes>



Build failed in Jenkins: Hive-trunk-h0.21 #886

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Hive-trunk-h0.21/886/>

------------------------------------------
[...truncated 31706 lines...]
    [junit] OK
    [junit] PREHOOK: query: select count(1) as cnt from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-45_847_6699736544887546667/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks determined at compile time: 1
    [junit] In order to change the average load for a reducer (in bytes):
    [junit]   set hive.exec.reducers.bytes.per.reducer=<number>
    [junit] In order to limit the maximum number of reducers:
    [junit]   set hive.exec.reducers.max=<number>
    [junit] In order to set a constant number of reducers:
    [junit]   set mapred.reduce.tasks=<number>
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-10 14:09:49,074 null map = 100%,  reduce = 100%
    [junit] Ended Job = job_local_0001
    [junit] POSTHOOK: query: select count(1) as cnt from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-45_847_6699736544887546667/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101409_1179642052.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (num int)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (num int)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'> into table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] Copying data from <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table default.testhivedrivertable
    [junit] POSTHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'> into table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: select * from testhivedrivertable limit 10
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-50_621_3509909480508959253/-mr-10000
    [junit] POSTHOOK: query: select * from testhivedrivertable limit 10
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-50_621_3509909480508959253/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101409_90257593.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (num int)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (num int)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101409_751941123.txt>
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101409_800999725.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'> into table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] Copying data from <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table default.testhivedrivertable
    [junit] POSTHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'> into table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-51_820_5340452372770500098/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-10 14:09:54,428 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] POSTHOOK: query: select key, value from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-51_820_5340452372770500098/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-54_586_6419728267065200525/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-10 14:09:57,200 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] POSTHOOK: query: select key, value from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-54_586_6419728267065200525/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-58_246_7987841676957405491/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-10 14:10:00,856 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] POSTHOOK: query: select key, value from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-09-58_246_7987841676957405491/-mr-10000
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101410_781635653.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'> into table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] Copying data from <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table default.testhivedrivertable
    [junit] POSTHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'> into table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable where key > 10
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-10-01_690_3720320864848006988/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-10 14:10:04,447 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] POSTHOOK: query: select key, value from testhivedrivertable where key > 10
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-10-01_690_3720320864848006988/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: select count(1) as c from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-10-04_643_1047391807830474963/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks determined at compile time: 1
    [junit] In order to change the average load for a reducer (in bytes):
    [junit]   set hive.exec.reducers.bytes.per.reducer=<number>
    [junit] In order to limit the maximum number of reducers:
    [junit]   set hive.exec.reducers.max=<number>
    [junit] In order to set a constant number of reducers:
    [junit]   set mapred.reduce.tasks=<number>
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-10 14:10:07,393 null map = 100%,  reduce = 100%
    [junit] Ended Job = job_local_0001
    [junit] POSTHOOK: query: select count(1) as c from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-08-10_14-10-04_643_1047391807830474963/-mr-10000
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101410_1728842671.txt>
    [junit] <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/service/../data/files/invalidpath/sample.jar> does not exist
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101410_444839628.txt>
    [junit] <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/service/../data/files/invalidpath/sample.txt> does not exist
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build/service/tmp/hive_job_log_hudson_201108101410_421209815.txt>
    [junit] <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/service/../data/files/invalidpath/sample.zip> does not exist
    [junit] ------------- ---------------- ---------------
    [junit] 
    [junit] Testcase: testExecute took 10.255 sec
    [junit] Testcase: testNonHiveCommand took 0.826 sec
    [junit] Testcase: testMetastore took 0.338 sec
    [junit] Testcase: testGetClusterStatus took 0.105 sec
    [junit] Testcase: testFetch took 9.771 sec
    [junit] Testcase: testDynamicSerde took 7.334 sec
    [junit] Testcase: testAddJarShouldFailIfJarNotExist took 0.044 sec
    [junit] Testcase: testAddFileShouldFailIfFileNotExist took 0.042 sec
    [junit] Testcase: testAddArchiveShouldFailIfFileNotExist took 0.042 sec

BUILD FAILED
<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build.xml>:208: Keepgoing execution: 1 of 10 iterations failed.

Total time: 307 minutes 3 seconds
Archiving artifacts
Recording test results