You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@tajo.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2015/08/16 04:26:20 UTC
Build failed in Jenkins: Tajo-master-jdk8-nightly #107
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/107/>
------------------------------------------
[...truncated 733285 lines...]
2015-08-16 02:26:42,110 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-16 02:26:42,110 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-16 02:26:42,114 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34596 is added to blk_1073748705_7881{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-c88053b7-df50-4797-9186-131c6a712883:NORMAL:127.0.0.1:34596|FINALIZED]]} size 0
2015-08-16 02:26:42,114 INFO: org.apache.tajo.worker.TaskAttemptContext (setState(142)) - Query status of ta_1439690731887_2052_000001_000000_00 is changed to TA_SUCCEEDED
2015-08-16 02:26:42,114 INFO: org.apache.tajo.worker.TaskImpl (run(460)) - ta_1439690731887_2052_000001_000000_00 completed. Worker's task counter - total:1, succeeded: 1, killed: 0, failed: 0
2015-08-16 02:26:42,115 INFO: org.apache.tajo.querymaster.Stage (transition(1298)) - Stage - eb_1439690731887_2052_000001 finalize NONE_SHUFFLE (total=1, success=1, killed=0)
2015-08-16 02:26:42,115 INFO: org.apache.tajo.querymaster.DefaultTaskScheduler (stop(160)) - Task Scheduler stopped
2015-08-16 02:26:42,115 INFO: org.apache.tajo.querymaster.DefaultTaskScheduler (run(122)) - TaskScheduler schedulingThread stopped
2015-08-16 02:26:42,115 INFO: org.apache.tajo.querymaster.Stage (transition(1356)) - Stage completed - eb_1439690731887_2052_000001 (total=1, success=1, killed=0)
2015-08-16 02:26:42,115 INFO: org.apache.tajo.querymaster.Query (handle(855)) - Processing q_1439690731887_2052 of type STAGE_COMPLETED
2015-08-16 02:26:42,115 INFO: org.apache.tajo.engine.planner.global.ParallelExecutionQueue (next(95)) - Next executable block eb_1439690731887_2052_000002
2015-08-16 02:26:42,115 INFO: org.apache.tajo.querymaster.Query (transition(802)) - Complete Stage[eb_1439690731887_2052_000001], State: SUCCEEDED, 1/1.
2015-08-16 02:26:42,116 INFO: org.apache.tajo.querymaster.Query (handle(855)) - Processing q_1439690731887_2052 of type QUERY_COMPLETED
2015-08-16 02:26:42,116 INFO: org.apache.tajo.worker.TaskManager (stopExecutionBlock(161)) - Stopped execution block:eb_1439690731887_2052_000001
2015-08-16 02:26:42,116 INFO: org.apache.tajo.querymaster.Query (finalizeQuery(528)) - Can't find partitions for adding.
2015-08-16 02:26:42,116 INFO: org.apache.tajo.querymaster.Query (handle(873)) - q_1439690731887_2052 Query Transitioned from QUERY_RUNNING to QUERY_SUCCEEDED
2015-08-16 02:26:42,116 INFO: org.apache.tajo.querymaster.QueryMasterTask (handle(295)) - Query completion notified from q_1439690731887_2052 final state: QUERY_SUCCEEDED
2015-08-16 02:26:42,117 INFO: org.apache.tajo.master.QueryInProgress (heartbeat(252)) - Received QueryMaster heartbeat:q_1439690731887_2052,state=QUERY_SUCCEEDED,progress=1.0, queryMaster=asf900.gq1.ygridcore.net
2015-08-16 02:26:42,117 INFO: org.apache.tajo.master.QueryManager (stopQuery(275)) - Stop QueryInProgress:q_1439690731887_2052
2015-08-16 02:26:42,117 INFO: org.apache.tajo.querymaster.QueryMasterTask (serviceStop(172)) - Stopping QueryMasterTask:q_1439690731887_2052
2015-08-16 02:26:42,117 INFO: org.apache.tajo.master.QueryInProgress (stopProgress(117)) - =========================================================
2015-08-16 02:26:42,117 INFO: org.apache.tajo.master.QueryInProgress (stopProgress(118)) - Stop query:q_1439690731887_2052
2015-08-16 02:26:42,117 INFO: org.apache.tajo.querymaster.QueryMasterTask (cleanupQuery(471)) - Cleanup resources of all workers. Query: q_1439690731887_2052, workers: 1
2015-08-16 02:26:42,118 INFO: org.apache.tajo.querymaster.QueryMasterTask (serviceStop(188)) - Stopped QueryMasterTask:q_1439690731887_2052
2015-08-16 02:26:42,333 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-16 02:26:42,334 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-16 02:26:42,335 INFO: org.apache.tajo.master.TajoMasterClientService (getQueryResultData(579)) - Send result to client for 6727fcca-1d43-4888-9ab3-a22fe8fb2108,q_1439690731887_2052, 2 rows
2015-08-16 02:26:42,336 INFO: org.apache.tajo.master.TajoMasterClientService (getQueryResultData(579)) - Send result to client for 6727fcca-1d43-4888-9ab3-a22fe8fb2108,q_1439690731887_2052, 0 rows
2015-08-16 02:26:42,338 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 6727fcca-1d43-4888-9ab3-a22fe8fb2108 is removed.
2015-08-16 02:26:42,339 INFO: org.apache.tajo.master.GlobalEngine (updateQuery(237)) - SQL: DROP TABLE IF EXISTS "TestTajoJdbc".table1
2015-08-16 02:26:42,339 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(280)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-16 02:26:42,340 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - =============================================
2015-08-16 02:26:42,340 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(283)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-16 02:26:42,340 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-16 02:26:42,340 INFO: org.apache.tajo.master.exec.DDLExecutor (dropTable(310)) - relation "TestTajoJdbc.table1" is already exists.
2015-08-16 02:26:42,341 INFO: org.apache.tajo.master.GlobalEngine (updateQuery(237)) - SQL: DROP TABLE IF EXISTS testaltertablepartition
2015-08-16 02:26:42,341 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(280)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-16 02:26:42,341 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - =============================================
2015-08-16 02:26:42,341 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(283)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-16 02:26:42,341 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-16 02:26:42,342 INFO: org.apache.tajo.catalog.CatalogServer (dropTable(697)) - relation "TestTajoJdbc.testaltertablepartition" is deleted from the catalog (127.0.0.1:20683)
2015-08-16 02:26:42,342 INFO: org.apache.tajo.master.exec.DDLExecutor (dropTable(327)) - relation "TestTajoJdbc.testaltertablepartition" is dropped.
2015-08-16 02:26:42,343 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 3361986d-4e7b-4ec1-91b9-aba9ea1764b8 is removed.
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.543 sec - in org.apache.tajo.jdbc.TestTajoJdbc
2015-08-16 02:26:42,348 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-16 02:26:42,351 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-16 02:26:42,351 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
eader initialized will read a total of 2 records.
Aug 16, 2015 2:08:48 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:08:48 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 2
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:09:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 16, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 212
Aug 16, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 5 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN_DICTIONARY], dic { 3 entries, 12B raw, 3B comp}
Aug 16, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate_function] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 16, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 16, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 16, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 5 records.
Aug 16, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 16, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 5
2015-08-16 02:26:42,356 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
2015-08-16 02:26:42,358 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-16 02:26:42,371 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-16 02:26:42,371 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-16 02:26:42,371 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-16 02:26:42,372 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf900.gq1.ygridcore.net_20687 stopped.
2015-08-16 02:26:42,373 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-16 02:26:42,375 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:20686) shutdown
2015-08-16 02:26:42,375 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-16 02:26:42,375 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:20683) shutdown
2015-08-16 02:26:42,375 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:20683) shutdown
2015-08-16 02:26:42,378 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_20685 stopped.
2015-08-16 02:26:42,384 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34596 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-19571bff-196c-426d-b6e5-e183147c41bf:NORMAL:127.0.0.1:34596|RBW]]} size 704
2015-08-16 02:26:42,385 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-16 02:26:42,387 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:20685) shutdown
2015-08-16 02:26:42,387 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:20684) shutdown
2015-08-16 02:26:42,396 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:20682) shutdown
2015-08-16 02:26:42,396 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-16 02:26:42,401 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:20684)]: ConnectException: Connection refused: /127.0.0.1:20684 Try to reconnect : /127.0.0.1:20684
2015-08-16 02:26:42,426 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-16 02:26:42,426 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-16 02:26:42,427 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:20689) shutdown
2015-08-16 02:26:42,428 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-16 02:26:42,429 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-16 02:26:42,431 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-16 02:26:42,431 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-16 02:26:42,432 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:20688) shutdown
2015-08-16 02:26:42,432 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-16 02:26:42,432 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:20687) shutdown
2015-08-16 02:26:42,432 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-16 02:26:42,433 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
2015-08-16 02:26:43,402 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:20684)]: ConnectException: Connection refused: /127.0.0.1:20684 Try to reconnect : /127.0.0.1:20684
Results :
Failed tests:
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1691, Failures: 3, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.546 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.104 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.558 s]
[INFO] Tajo Common ....................................... SUCCESS [ 21.833 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.205 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.817 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.103 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.202 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [01:26 min]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.369 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.698 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 7.940 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 43.625 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 0.982 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.489 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.640 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 2.477 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.829 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.477 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.367 s]
[INFO] Tajo Core ......................................... SUCCESS [ 8.124 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.914 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 10.396 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.044 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.963 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.211 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.929 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.339 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.128 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:32 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:29 min
[INFO] Finished at: 2015-08-16T02:26:44+00:00
[INFO] Final Memory: 141M/1931M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58863387 bytes
Compression is 0.0%
Took 19 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #123
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/123/changes>
Changes:
[jihoonson] TAJO-1811: Catalog server address must be set dynamically during test.
------------------------------------------
[...truncated 722228 lines...]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1718, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.628 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.363 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.645 s]
[INFO] Tajo Common ....................................... SUCCESS [ 27.385 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.187 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.680 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.244 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.197 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.307 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.371 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.163 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 2.453 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 43.749 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 0.898 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.549 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.680 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.840 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.488 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.335 s]
[INFO] Tajo Core ......................................... SUCCESS [ 8.756 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.913 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 9.773 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.041 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.950 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.002 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.431 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.222 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 31.237 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.944 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.306 s]
[INFO] Tajo Core Tests ................................... FAILURE [20:35 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:22 min
[INFO] Finished at: 2015-09-03T14:43:43+00:00
[INFO] Final Memory: 145M/1775M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59312758 bytes
Compression is 0.0%
Took 15 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #122
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/122/changes>
Changes:
[jhkim] TAJO-1810: Remove QueryMasterTask cache immediately, if it stored to persistent storage.
[jhkim] TAJO-1738: Improve off-heap RowBlock.
[hyunsik] TAJO-1812: Timezone support in JSON file format.
[hyunsik] TAJO-1749: Refine JDBC exceptions to better handle exceptional cases.
------------------------------------------
[...truncated 722180 lines...]
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:587 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1718, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.908 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.836 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.817 s]
[INFO] Tajo Common ....................................... SUCCESS [ 29.406 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.452 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.628 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.440 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.244 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.603 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.426 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 10.034 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 2.767 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 45.450 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.078 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.720 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.017 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.922 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 4.162 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.460 s]
[INFO] Tajo Core ......................................... SUCCESS [ 9.944 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.981 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 12.948 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.069 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.006 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.172 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 5.472 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.360 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 32.226 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 1.005 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.461 s]
[INFO] Tajo Core Tests ................................... FAILURE [20:48 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:51 min
[INFO] Finished at: 2015-09-03T13:44:27+00:00
[INFO] Final Memory: 146M/2149M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59322424 bytes
Compression is 0.0%
Took 15 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #121
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/121/changes>
Changes:
[jhkim] TAJO-993: Cleanup the result data in HDFS after query finished.
------------------------------------------
[...truncated 718910 lines...]
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
Tests run: 1713, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.832 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.738 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 3.058 s]
[INFO] Tajo Common ....................................... SUCCESS [ 22.477 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.558 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.571 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.589 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.292 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.547 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.431 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.791 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.659 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 50.458 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.004 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.718 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.275 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 2.004 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 4.753 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.404 s]
[INFO] Tajo Core ......................................... SUCCESS [ 10.342 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.936 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.534 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.081 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.963 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.061 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.842 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.488 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 26.030 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.982 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.459 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:04 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:03 min
[INFO] Finished at: 2015-09-03T03:15:26+00:00
[INFO] Final Memory: 145M/1757M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59291312 bytes
Compression is 0.0%
Took 14 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #120
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/120/changes>
Changes:
[jihoonson] TAJO-1802: PythonScriptEngine copies controller and tajo util whenever it is initialized.
------------------------------------------
[...truncated 718590 lines...]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1711, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.560 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.075 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.396 s]
[INFO] Tajo Common ....................................... SUCCESS [ 22.857 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.176 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.780 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.425 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.191 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 45.260 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.432 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.176 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.099 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 42.832 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 0.997 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.609 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.689 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.840 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.566 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.370 s]
[INFO] Tajo Core ......................................... SUCCESS [ 8.649 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.917 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 9.884 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.038 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.022 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 0.992 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.532 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.519 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 25.997 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.933 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.345 s]
[INFO] Tajo Core Tests ................................... FAILURE [20:21 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:03 min
[INFO] Finished at: 2015-09-02T10:44:31+00:00
[INFO] Final Memory: 142M/1896M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59288874 bytes
Compression is 0.0%
Took 16 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #119
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/119/changes>
Changes:
[jhkim] TAJO-1707: Rack local count can be more than actual number of tasks.
------------------------------------------
[...truncated 718510 lines...]
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1711, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.699 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.506 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.637 s]
[INFO] Tajo Common ....................................... SUCCESS [ 23.316 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.588 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.725 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.702 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.194 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.378 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.370 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.059 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.111 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [01:00 min]
[INFO] Tajo PullServer ................................... SUCCESS [ 0.979 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.611 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.837 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.910 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.528 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.372 s]
[INFO] Tajo Core ......................................... SUCCESS [ 10.011 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.933 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 10.157 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.075 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.971 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.043 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.783 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.302 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 26.230 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.969 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.707 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:23 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:28 min
[INFO] Finished at: 2015-09-02T08:41:46+00:00
[INFO] Final Memory: 140M/1695M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59297293 bytes
Compression is 0.0%
Took 20 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #118
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/118/changes>
Changes:
[jihoonson] TAJO-1808: Wrong table type problem in catalog.
------------------------------------------
[...truncated 718479 lines...]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1711, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.591 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.489 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.739 s]
[INFO] Tajo Common ....................................... SUCCESS [ 22.617 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.216 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.733 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.195 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.206 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.402 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.340 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.421 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.337 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 41.907 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 0.842 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.597 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.791 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.963 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.825 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.387 s]
[INFO] Tajo Core ......................................... SUCCESS [ 9.358 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.942 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.031 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.064 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.953 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 0.995 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.558 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.101 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 26.340 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.982 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.244 s]
[INFO] Tajo Core Tests ................................... FAILURE [20:41 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:26 min
[INFO] Finished at: 2015-09-02T06:03:21+00:00
[INFO] Final Memory: 148M/1997M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59287483 bytes
Compression is 0.0%
Took 15 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #117
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/117/changes>
Changes:
[jihoonson] TAJO-1803: Use in-memory derby as the default catalog for unit tests.
------------------------------------------
[...truncated 718557 lines...]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1711, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.872 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.883 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 3.309 s]
[INFO] Tajo Common ....................................... SUCCESS [ 23.589 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.445 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.862 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.933 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.344 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.526 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.388 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 10.061 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.717 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 44.019 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.075 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.701 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.233 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 2.287 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.872 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.426 s]
[INFO] Tajo Core ......................................... SUCCESS [ 10.275 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.953 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.186 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.076 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.960 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.090 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 4.358 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.485 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 27.091 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 1.004 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.599 s]
[INFO] Tajo Core Tests ................................... FAILURE [20:49 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:46 min
[INFO] Finished at: 2015-09-01T10:40:42+00:00
[INFO] Final Memory: 140M/1718M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 59286862 bytes
Compression is 0.0%
Took 15 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #116
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/116/changes>
Changes:
[jihoonson] TAJO-1801: Table name is not unique of tableDescMap in QueryMasterTask.
------------------------------------------
[...truncated 717024 lines...]
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1711, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.623 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.483 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 3.007 s]
[INFO] Tajo Common ....................................... SUCCESS [ 23.766 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.233 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.703 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.632 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.186 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.628 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.343 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.626 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.287 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 46.308 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.038 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.647 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.721 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 2.022 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.569 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.438 s]
[INFO] Tajo Core ......................................... SUCCESS [ 9.307 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.913 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 10.068 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.039 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.972 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.007 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.401 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.007 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 27.876 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.963 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.592 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:03 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:55 min
[INFO] Finished at: 2015-08-27T17:40:32+00:00
[INFO] Final Memory: 143M/1699M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58893301 bytes
Compression is 0.0%
Took 15 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #115
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/115/changes>
Changes:
[blrunner] TAJO-1798: Dynamic partitioning occasionally fails.
------------------------------------------
[...truncated 716669 lines...]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testWhereClauseJoin6:98->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestInnerJoinQuery.testJoinWithMultipleJoinQual1:112->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.p
2: type=Broadcast, tables=default.s]
3: type=Broadcast,...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.ps
2: type=Broadcast, tables=default.p]
3: type=Broadcast,...>
TestInnerJoinQuery.testBroadcastTwoPartJoin:266->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[orders
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.lineitem]
3: type=Broadcast,...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.part
2: type=Broadcast, tables=default.orders]
3: type=Broadcast,...>
TestInnerJoinQuery.testInnerJoinWithThetaJoinConditionInWhere:56->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
SELECTION(6)
=> ...> but was:<...ast, tables=default.[b]
SELECTION(6)
=> ...>
TestInnerJoinQuery.testDifferentTypesJoinCondition:209->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4]
JOIN(5)(INNER)
=...> but was:<...st, tables=default.t[3]
JOIN(5)(INNER)
=...>
TestInnerJoinQuery.testTPCHQ2Join:105->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...> but was:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.part]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestCrossJoin.testCrossJoinOfOneSmallTableAndJoin:152->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[supplier
1: type=Broadcast, tables=default.region]
2: type=Broadcast,...> but was:<...ast, tables=default.[region
1: type=Broadcast, tables=default.supplier]
2: type=Broadcast,...>
TestCrossJoin.testCrossJoinWithThetaJoinConditionInWhere:117->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[a]
PROJECTION(3)
=>...> but was:<...ast, tables=default.[b]
PROJECTION(3)
=>...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests in error:
TestJoinOnPartitionedTables.testPartialFilterPushDown:95->QueryTestCaseBase.runSimpleTests:596 ยป UndefinedTable
Tests run: 1711, Failures: 45, Errors: 1, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 2.141 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 2.024 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 3.308 s]
[INFO] Tajo Common ....................................... SUCCESS [ 25.191 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.586 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.873 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 7.316 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.338 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.729 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.423 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 10.165 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.367 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [01:03 min]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.127 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.678 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.088 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.948 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.972 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.606 s]
[INFO] Tajo Core ......................................... SUCCESS [ 9.730 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.943 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.525 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.065 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.016 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.057 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 4.169 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.646 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 25.577 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 1.038 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.341 s]
[INFO] Tajo Core Tests ................................... FAILURE [19:57 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:13 min
[INFO] Finished at: 2015-08-26T09:40:14+00:00
[INFO] Final Memory: 145M/1778M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58889781 bytes
Compression is 0.0%
Took 14 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #114
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/114/changes>
Changes:
[jihoonson] TAJO-1766: Improve the performance of cross join.
------------------------------------------
[...truncated 730689 lines...]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase1:361->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase2:379->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t1]
JOIN(9)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t4]
JOIN(9)(RIGHT_OUTE...>
TestOuterJoinQuery.testRightOuterJoinPredicationCaseByCase3:393->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t1]
JOIN(8)(RIGHT_OUTE...> but was:<...st, tables=default.t[1
1: type=Broadcast, tables=default.t2]
JOIN(8)(RIGHT_OUTE...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2:249->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t2]
2: type=Broadcast,...> but was:<...st, tables=default.t[2
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...>
TestOuterJoinQuery.testOuterJoinAndCaseWhen1:132->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
SORT(15)
=> Sort...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
SORT(15)
=> Sort...>
TestOuterJoinQuery.testLeftOuterJoinPredicationCaseByCase2_1:267->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.t[4
1: type=Broadcast, tables=default.t3]
2: type=Broadcast,...> but was:<...st, tables=default.t[3
1: type=Broadcast, tables=default.t4]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin2:424->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[b
1: type=Broadcast, tables=default.c]
2: type=Broadcast,...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.b]
2: type=Broadcast,...>
TestOuterJoinQuery.testLeftOuterJoin3:432->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[e
1: type=Broadcast, tables=default.b
2: type=Broadcast, tables=default.c
3: type=Broadcast, tables=default.d]
4: type=Broadcast,...> but was:<...ast, tables=default.[d
1: type=Broadcast, tables=default.c
2: type=Broadcast, tables=default.b
3: type=Broadcast, tables=default.e]
4: type=Broadcast,...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestOuterJoinWithSubQuery.testLeftOuterJoinWithEmptySubquery1:87->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[jointable12
1: type=Broadcast, tables=default.lineitem]
SORT(17)
=> Sort...> but was:<...ast, tables=default.[lineitem
1: type=Broadcast, tables=default.jointable12]
SORT(17)
=> Sort...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual3:63->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testJoinWithMultipleJoinQual4:70->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[ps
1: type=Broadcast, tables=default.]s
2: type=Broadcast...> but was:<...ast, tables=default.[s
1: type=Broadcast, tables=default.p]s
2: type=Broadcast...>
TestInnerJoinWithSubQuery.testComplexJoinCondition5:95->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...st, tables=default.n[ation]
SORT(12)
=> Sort...> but was:<...st, tables=default.n[1]
SORT(12)
=> Sort...>
TestInnerJoinWithSubQuery.testBroadcastSubquery2:123->QueryTestCaseBase.runSimpleTests:585 Plan Verification for: 1 th test expected:<...ast, tables=default.[part
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.c]
GROUP_BY(36)()
=...> but was:<...ast, tables=default.[c
1: type=Broadcast, tables=default.a
2: type=Broadcast, tables=default.part]
GROUP_BY(36)()
=...>
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1709, Failures: 45, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 2.071 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.681 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 4.113 s]
[INFO] Tajo Common ....................................... SUCCESS [ 25.811 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.974 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.659 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.403 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 3.178 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.050 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.502 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 10.205 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 11.325 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 54.918 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 4.188 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.518 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.502 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.746 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 4.054 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.338 s]
[INFO] Tajo Core ......................................... SUCCESS [ 9.719 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.921 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 12.238 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.070 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.980 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.054 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 5.353 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 4.249 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 25.501 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.960 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.388 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:06 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:24 min
[INFO] Finished at: 2015-08-25T10:43:08+00:00
[INFO] Final Memory: 154M/2552M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58885867 bytes
Compression is 0.0%
Took 14 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #113
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/113/changes>
Changes:
[jhkim] TAJO-1799: Fix incorrect event handler when kill-query failed.
------------------------------------------
[...truncated 727905 lines...]
2015-08-24 04:45:38,938 INFO: org.apache.tajo.pullserver.TajoPullServerService (init(203)) - Tajo PullServer initialized: readaheadLength=4194304
2015-08-24 04:45:38,944 INFO: org.apache.tajo.worker.Fetcher (get(135)) - Get real fetch from remote host
2015-08-24 04:45:38,947 INFO: org.apache.tajo.worker.Fetcher (get(174)) - Fetcher finished:3 ms, FETCH_FAILED, URI:http://127.0.0.1:36460/?qid=q_0000000000000_0000&sid=1&p=1&type=h&ta=1_0
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.949 sec - in org.apache.tajo.worker.TestFetcher
Running org.apache.tajo.worker.TestTaskManager
2015-08-24 04:45:38,983 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:38,984 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-24 04:45:38,984 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:38,984 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:38,989 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:1536, Disks:4, vCores:3)
2015-08-24 04:45:38,991 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:38,993 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:38,993 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:38,993 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-24 04:45:39,036 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:39,038 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-24 04:45:39,038 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:39,038 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:39,042 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:39,044 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:39,046 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:39,046 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:39,047 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.103 sec - in org.apache.tajo.worker.TestTaskManager
Running org.apache.tajo.worker.TestNodeStatusUpdater
2015-08-24 04:45:39,091 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1000, Disks:2, vCores:2)
2015-08-24 04:45:39,094 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:39,094 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:39,096 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:39,097 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:39,138 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1000, Disks:2, vCores:2)
2015-08-24 04:45:39,145 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:39,146 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:40,150 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:40,152 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:40,192 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1000, Disks:2, vCores:2)
2015-08-24 04:45:40,195 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:40,196 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:40,197 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:40,199 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:40,243 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1000, Disks:2, vCores:2)
2015-08-24 04:45:40,246 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:40,246 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:40,248 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:40,250 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.201 sec - in org.apache.tajo.worker.TestNodeStatusUpdater
Running org.apache.tajo.worker.TestDeletionService
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.073 sec - in org.apache.tajo.worker.TestDeletionService
Running org.apache.tajo.worker.TestNodeResourceManager
2015-08-24 04:45:42,376 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:42,377 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-24 04:45:42,378 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:42,378 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:42,653 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:1536, Disks:4, vCores:3)
2015-08-24 04:45:46,362 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:46,362 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:46,366 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:46,383 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-24 04:45:46,427 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:46,429 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-24 04:45:46,429 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:46,432 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:46,435 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:46,437 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:46,437 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:46,437 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-24 04:45:46,507 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:46,550 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-24 04:45:46,550 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:46,551 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:46,567 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:46,567 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:46,568 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:46,572 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-24 04:45:46,632 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-24 04:45:46,644 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-24 04:45:46,644 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-24 04:45:46,644 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-24 04:45:46,646 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-24 04:45:46,647 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-24 04:45:46,647 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:46,647 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.324 sec - in org.apache.tajo.worker.TestNodeResourceManager
eader initialized will read a total of 2 records.
Aug 24, 2015 4:25:56 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:25:56 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 2 ms. row count = 2
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:19 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:26:20 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 212
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 5 values, 10B raw, 10B comp, 1 pages, encodings: [RLE, PLAIN_DICTIONARY, BIT_PACKED], dic { 3 entries, 12B raw, 3B comp}
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate_function] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [RLE, BIT_PACKED, PLAIN]
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 5 records.
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 24, 2015 4:26:23 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 2 ms. row count = 5
2015-08-24 04:45:46,736 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
2015-08-24 04:45:46,736 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-24 04:45:46,676 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-24 04:45:46,679 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-24 04:45:46,764 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-24 04:45:46,764 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-24 04:45:46,764 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-24 04:45:46,764 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
2015-08-24 04:45:46,764 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 6f750fb9-3326-4cc8-8959-ca8baf0c8fc6 is removed.
2015-08-24 04:45:46,764 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 0c76004b-1899-4470-bf3c-5e3afa3c7463 is removed.
2015-08-24 04:45:46,790 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:15015) shutdown
2015-08-24 04:45:46,792 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-24 04:45:46,796 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:15012) shutdown
2015-08-24 04:45:46,798 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:15012) shutdown
2015-08-24 04:45:46,801 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_15014 stopped.
2015-08-24 04:45:46,807 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf903.gq1.ygridcore.net_15016 stopped.
2015-08-24 04:45:46,807 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-24 04:45:46,844 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:42058 is added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-1ab2f43a-047a-456e-b9f9-3d8cf7bc97eb:NORMAL:127.0.0.1:42058|RBW]]} size 1960
2015-08-24 04:45:46,845 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-24 04:45:46,845 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:15014) shutdown
2015-08-24 04:45:46,856 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:15013) shutdown
2015-08-24 04:45:46,870 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:15011) shutdown
2015-08-24 04:45:46,871 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-24 04:45:46,901 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-24 04:45:46,901 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-24 04:45:46,911 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:15018) shutdown
2015-08-24 04:45:46,912 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-24 04:45:46,914 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-24 04:45:46,917 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-24 04:45:46,917 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-24 04:45:46,918 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:15017) shutdown
2015-08-24 04:45:46,918 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-24 04:45:46,925 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:15016) shutdown
2015-08-24 04:45:46,925 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-24 04:45:46,928 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
Results :
Failed tests:
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1689, Failures: 3, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 3.974 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 5.172 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 5.518 s]
[INFO] Tajo Common ....................................... SUCCESS [ 29.146 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.731 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.985 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 7.074 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.539 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 49.359 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.709 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 10.730 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 10.012 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [01:19 min]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.764 s]
[INFO] Tajo Client ....................................... SUCCESS [ 3.381 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.863 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 2.397 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 5.480 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.396 s]
[INFO] Tajo Core ......................................... SUCCESS [ 14.355 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 1.001 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 15.647 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.099 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.022 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.112 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 5.391 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.884 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 34.634 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 1.024 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 6.711 s]
[INFO] Tajo Core Tests ................................... FAILURE [24:25 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 29:39 min
[INFO] Finished at: 2015-08-24T04:45:48+00:00
[INFO] Final Memory: 146M/1953M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58885222 bytes
Compression is 0.0%
Took 18 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #112
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/112/changes>
Changes:
[hyunsik] TAJO-1790: TestTajoClientV2::testExecuteQueryAsyncWithListener occasionally is failed.
[hyunsik] TAJO-1741: Two tables having same time zone display different timestamps.
[hyunsik] TAJO-1792: tajo-cluster-tests is not available when it is used as an external maven module.
[hyunsik] TAJO-1787: Remove unused and legacy exceptions.
[hyunsik] TAJO-1737: Implement SQL Parser rule for Map type.
[jhkim] TAJO-1596: TestPythonFunctions occasionally fails.
[jhkim] TAJO-1783: Query result is not returned by invalid output path.
------------------------------------------
[...truncated 727795 lines...]
2015-08-21 19:14:57,533 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:1000, Disks:3, vCores:2)
2015-08-21 19:14:57,533 INFO: org.apache.tajo.worker.TaskImpl (<init>(114)) - Task basedir is created (<https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/output/4)>
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskAttemptContext (setState(142)) - Query status of ta_1440183233682_2058_000004_000000_00 is changed to TA_PENDING
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskImpl (init(195)) - Initializing: ta_1440183233682_2058_000004_000000_00
2015-08-21 19:14:57,534 INFO: org.apache.tajo.storage.FileTablespace (getAppenderFilePath(294)) - Output File Path: hdfs://localhost:53245/tmp/tajo-jenkins/staging/q_1440183233682_2058/RESULT/part-04-000000-000
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskImpl (initPlan(145)) - Output File Path: hdfs://localhost:53245/tmp/tajo-jenkins/staging/q_1440183233682_2058/RESULT/part-04-000000-000
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskImpl (initPlan(150)) - ==================================
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskImpl (initPlan(151)) - * Stage ta_1440183233682_2058_000004_000000_00 is initialized
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskImpl (initPlan(152)) - * InterQuery: false, Fragments (num: 1), Fetches (total:1) :
2015-08-21 19:14:57,534 INFO: org.apache.tajo.worker.TaskImpl (initPlan(162)) - * Local task dir: <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/output/4/0_0>
2015-08-21 19:14:57,535 INFO: org.apache.tajo.worker.TaskImpl (initPlan(167)) - ==================================
2015-08-21 19:14:57,535 INFO: org.apache.tajo.worker.TaskImpl (init(214)) - the directory is created <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/in/eb_1440183233682_2058_000004/0/0/eb_1440183233682_2058_000003>
2015-08-21 19:14:57,536 INFO: org.apache.tajo.worker.TaskImpl (getFetchRunners(692)) - Try to get local file chunk at local host
2015-08-21 19:14:57,536 INFO: org.apache.tajo.worker.TaskImpl (getLocalStoredFileChunk(734)) - getLocalStoredFileChunk starts
2015-08-21 19:14:57,536 INFO: org.apache.tajo.worker.TaskImpl (getLocalStoredFileChunk(769)) - PullServer request param: shuffleType=r, sid=3, partId=0, taskIds=[0_0]
2015-08-21 19:14:57,560 INFO: org.apache.tajo.pullserver.TajoPullServerService (getFileChunks(686)) - BSTIndex is loaded from disk ((1,1), (5,NULL)
2015-08-21 19:14:57,560 INFO: org.apache.tajo.pullserver.TajoPullServerService (getFileChunks(710)) - GET Request for <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/output/3/0_0/output/output> (start=(1,1), end=(5,NULL), last=true)
2015-08-21 19:14:57,560 INFO: org.apache.tajo.pullserver.TajoPullServerService (getFileChunks(779)) - Retrieve File Chunk: (start=0, length=49, fromRemote=false, ebId=null) <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/output/3/0_0/output/output>
2015-08-21 19:14:57,560 INFO: org.apache.tajo.worker.TaskImpl (getFetchRunners(720)) - Create a new Fetcher with storeChunk: (start=0, length=49, fromRemote=false, ebId=eb_1440183233682_2058_000003) <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/output/3/0_0/output/output>
2015-08-21 19:14:57,561 INFO: org.apache.tajo.worker.TaskAttemptContext (setState(142)) - Query status of ta_1440183233682_2058_000004_000000_00 is changed to TA_RUNNING
2015-08-21 19:14:57,561 INFO: org.apache.tajo.worker.Fetcher (get(128)) - Get pseudo fetch from local host
2015-08-21 19:14:57,561 INFO: org.apache.tajo.worker.TaskImpl (run(616)) - Add a new FileChunk to local chunk list
2015-08-21 19:14:57,561 INFO: org.apache.tajo.worker.TaskImpl (waitForFetch(362)) - ta_1440183233682_2058_000004_000000_00 All fetches are done!
2015-08-21 19:14:57,562 INFO: org.apache.tajo.worker.TaskImpl (localizeFetchedData(572)) - One local chunk is added to listTablets
2015-08-21 19:14:57,564 INFO: org.apache.tajo.engine.planner.physical.ExternalSortExec (info(70)) - [ta_1440183233682_2058_000004_000000_00] Total merge time: 0 msec
2015-08-21 19:14:57,564 INFO: org.apache.tajo.engine.planner.physical.ExternalSortExec (close(796)) - Delete file: "fragment": {"id": "eb_1440183233682_2058_000003", "path": <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/test-data/a7846be0-d9e3-4d0b-9ba3-0240003da36b/tajo-localdir/q_1440183233682_2058/output/3/0_0/output/output",> "start": 0,"length": 49}
2015-08-21 19:14:57,567 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33628 is added to blk_1073748618_7794{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-52df8432-30ed-4935-b15c-399d06c34052:NORMAL:127.0.0.1:33628|FINALIZED]]} size 0
2015-08-21 19:14:57,567 INFO: org.apache.tajo.worker.TaskAttemptContext (setState(142)) - Query status of ta_1440183233682_2058_000004_000000_00 is changed to TA_SUCCEEDED
2015-08-21 19:14:57,568 INFO: org.apache.tajo.worker.TaskImpl (run(460)) - ta_1440183233682_2058_000004_000000_00 completed. Worker's task counter - total:1, succeeded: 1, killed: 0, failed: 0
2015-08-21 19:14:57,568 INFO: org.apache.tajo.querymaster.Stage (transition(1298)) - Stage - eb_1440183233682_2058_000004 finalize NONE_SHUFFLE (total=1, success=1, killed=0)
2015-08-21 19:14:57,568 INFO: org.apache.tajo.querymaster.DefaultTaskScheduler (stop(160)) - Task Scheduler stopped
2015-08-21 19:14:57,568 INFO: org.apache.tajo.querymaster.Stage (transition(1356)) - Stage completed - eb_1440183233682_2058_000004 (total=1, success=1, killed=0)
2015-08-21 19:14:57,568 INFO: org.apache.tajo.querymaster.Query (handle(855)) - Processing q_1440183233682_2058 of type STAGE_COMPLETED
2015-08-21 19:14:57,568 INFO: org.apache.tajo.engine.planner.global.ParallelExecutionQueue (next(95)) - Next executable block eb_1440183233682_2058_000005
2015-08-21 19:14:57,569 INFO: org.apache.tajo.querymaster.Query (transition(802)) - Complete Stage[eb_1440183233682_2058_000004], State: SUCCEEDED, 4/4.
2015-08-21 19:14:57,569 INFO: org.apache.tajo.querymaster.Query (handle(855)) - Processing q_1440183233682_2058 of type QUERY_COMPLETED
2015-08-21 19:14:57,568 INFO: org.apache.tajo.querymaster.DefaultTaskScheduler (run(122)) - TaskScheduler schedulingThread stopped
2015-08-21 19:14:57,569 INFO: org.apache.tajo.worker.TaskManager (stopExecutionBlock(161)) - Stopped execution block:eb_1440183233682_2058_000004
2015-08-21 19:14:57,569 INFO: org.apache.tajo.querymaster.Query (finalizeQuery(528)) - Can't find partitions for adding.
2015-08-21 19:14:57,570 INFO: org.apache.tajo.querymaster.Query (handle(873)) - q_1440183233682_2058 Query Transitioned from QUERY_RUNNING to QUERY_SUCCEEDED
2015-08-21 19:14:57,570 INFO: org.apache.tajo.querymaster.QueryMasterTask (handle(295)) - Query completion notified from q_1440183233682_2058 final state: QUERY_SUCCEEDED
2015-08-21 19:14:57,570 INFO: org.apache.tajo.master.QueryInProgress (heartbeat(252)) - Received QueryMaster heartbeat:q_1440183233682_2058,state=QUERY_SUCCEEDED,progress=1.0, queryMaster=asf906.gq1.ygridcore.net
2015-08-21 19:14:57,570 INFO: org.apache.tajo.master.QueryManager (stopQuery(276)) - Stop QueryInProgress:q_1440183233682_2058
2015-08-21 19:14:57,570 INFO: org.apache.tajo.master.QueryInProgress (stopProgress(117)) - =========================================================
2015-08-21 19:14:57,570 INFO: org.apache.tajo.master.QueryInProgress (stopProgress(118)) - Stop query:q_1440183233682_2058
2015-08-21 19:14:57,570 INFO: org.apache.tajo.querymaster.QueryMasterTask (serviceStop(172)) - Stopping QueryMasterTask:q_1440183233682_2058
2015-08-21 19:14:57,570 INFO: org.apache.tajo.querymaster.QueryMasterTask (cleanupQuery(471)) - Cleanup resources of all workers. Query: q_1440183233682_2058, workers: 1
2015-08-21 19:14:57,571 INFO: org.apache.tajo.querymaster.QueryMasterTask (serviceStop(188)) - Stopped QueryMasterTask:q_1440183233682_2058
2015-08-21 19:14:57,641 INFO: org.apache.tajo.storage.FileTablespace (listStatus(368)) - Total input paths to process : 1
2015-08-21 19:14:57,642 INFO: org.apache.tajo.storage.FileTablespace (getSplits(572)) - Total # of splits: 1
2015-08-21 19:14:57,644 INFO: org.apache.tajo.master.TajoMasterClientService (getQueryResultData(579)) - Send result to client for 19581a0d-9ead-4aa4-8500-ce63849fb449,q_1440183233682_2058, 5 rows
2015-08-21 19:14:57,645 INFO: org.apache.tajo.master.TajoMasterClientService (getQueryResultData(579)) - Send result to client for 19581a0d-9ead-4aa4-8500-ce63849fb449,q_1440183233682_2058, 0 rows
2015-08-21 19:14:57,646 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 19581a0d-9ead-4aa4-8500-ce63849fb449 is removed.
2015-08-21 19:14:57,649 INFO: org.apache.tajo.session.SessionManager (createSession(80)) - Session 5144d40f-e5dd-4c4d-9bce-ceafb3b26dcd is created.
2015-08-21 19:14:57,650 INFO: org.apache.tajo.master.GlobalEngine (executeQuery(177)) - Query: create table "TEST_DESC_TABLE" (col1 int4, col2 int4)
2015-08-21 19:14:57,651 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-21 19:14:57,651 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-21 19:14:57,651 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(285)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-21 19:14:57,652 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(286)) - =============================================
2015-08-21 19:14:57,653 INFO: org.apache.tajo.catalog.CatalogServer (createTable(667)) - relation "default.TEST_DESC_TABLE" is added to the catalog (127.0.0.1:16414)
2015-08-21 19:14:57,656 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 5144d40f-e5dd-4c4d-9bce-ceafb3b26dcd is removed.
Tests run: 21, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.442 sec - in org.apache.tajo.cli.tsql.TestTajoCli
2015-08-21 19:14:57,662 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-21 19:14:57,662 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-21 19:14:57,662 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-21 19:14:57,667 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 316635e5-b4bd-4786-a15c-7944cab1bb43 is removed.
2015-08-21 19:14:57,668 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:16417) shutdown
2015-08-21 19:14:57,669 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-21 19:14:57,669 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:16414) shutdown
2015-08-21 19:14:57,670 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
2015-08-21 19:14:57,671 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-21 19:14:57,673 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-21 19:14:57,673 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-21 19:14:57,673 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
eader initialized will read a total of 2 records.
Aug 21, 2015 7:02:43 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:02:43 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 2
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:57 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 0 ms. row count = 1
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 0 ms. row count = 1
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 21, 2015 7:02:58 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 212
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 5 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN_DICTIONARY, RLE], dic { 3 entries, 12B raw, 3B comp}
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate_function] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [BIT_PACKED, PLAIN, RLE]
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 5 records.
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 21, 2015 7:03:01 PM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 5
2015-08-21 19:14:57,678 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:16414) shutdown
2015-08-21 19:14:57,680 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf906.gq1.ygridcore.net_16418 stopped.
2015-08-21 19:14:57,680 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-21 19:14:57,681 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_16416 stopped.
2015-08-21 19:14:57,684 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-21 19:14:57,685 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-21 19:14:57,688 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:16420) shutdown
2015-08-21 19:14:57,688 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-21 19:14:57,688 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33628 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-52df8432-30ed-4935-b15c-399d06c34052:NORMAL:127.0.0.1:33628|RBW]]} size 524
2015-08-21 19:14:57,692 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-21 19:14:57,690 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-21 19:14:57,693 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-21 19:14:57,693 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-21 19:14:57,695 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:16416) shutdown
2015-08-21 19:14:57,695 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:16415) shutdown
2015-08-21 19:14:57,695 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:16419) shutdown
2015-08-21 19:14:57,697 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-21 19:14:57,700 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:16413) shutdown
2015-08-21 19:14:57,700 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:16418) shutdown
2015-08-21 19:14:57,700 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-21 19:14:57,700 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-21 19:14:57,701 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
2015-08-21 19:14:57,713 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:16415)]: ClosedChannelException: Try to reconnect : /127.0.0.1:16415
2015-08-21 19:14:58,714 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:16415)]: ClosedChannelException: Try to reconnect : /127.0.0.1:16415
Results :
Failed tests:
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1689, Failures: 3, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 2.003 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 2.535 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 4.299 s]
[INFO] Tajo Common ....................................... SUCCESS [ 23.228 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.414 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.802 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.637 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.331 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [ 46.925 s]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.519 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.760 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.593 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 45.338 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.027 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.542 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.057 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.977 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.759 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.436 s]
[INFO] Tajo Core ......................................... SUCCESS [ 9.932 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.967 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.230 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.104 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.031 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.045 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.796 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.354 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 24.667 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.995 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.368 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:16 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 25:10 min
[INFO] Finished at: 2015-08-21T19:15:00+00:00
[INFO] Final Memory: 150M/2250M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58864764 bytes
Compression is 0.0%
Took 21 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #111
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/111/changes>
Changes:
[hyunsik] TAJO-1757: Add examples for TajoClient v2.
[hyunsik] TAJO-1745: Add positive and negative test methods.
[blrunner] TAJO-1775: HCatalogStore need to be deprecated.
[hyunsik] TAJO-1758: Some TajoRuntimeException should be restored as TajoException in client side.
[blrunner] TAJO-1775: HCatalogStore need to be deprecated. (missing change)
[jihoonson] TAJO-1781: Join condition is still not found when it exists in OR clause.
[jhkim] TAJO-1776: Fix Invalid column type in JDBC.
------------------------------------------
[...truncated 725143 lines...]
2015-08-20 05:01:00,146 INFO: org.apache.tajo.worker.Fetcher (get(135)) - Get real fetch from remote host
2015-08-20 05:01:00,149 INFO: org.apache.tajo.worker.Fetcher (get(174)) - Fetcher finished:3 ms, FETCH_FAILED, URI:http://127.0.0.1:35218/?qid=q_0000000000000_0000&sid=1&p=1&type=h&ta=1_0
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.835 sec - in org.apache.tajo.worker.TestFetcher
Running org.apache.tajo.worker.TestTaskManager
2015-08-20 05:01:00,190 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:00,192 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-20 05:01:00,192 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:00,192 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:00,197 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:1536, Disks:4, vCores:3)
2015-08-20 05:01:00,201 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:00,202 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:00,203 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:00,203 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-20 05:01:00,239 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:00,241 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-20 05:01:00,241 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:00,241 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:00,245 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:00,247 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:00,247 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:00,250 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:00,250 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.097 sec - in org.apache.tajo.worker.TestTaskManager
Running org.apache.tajo.worker.TestNodeStatusUpdater
2015-08-20 05:01:00,289 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1011, Disks:2, vCores:2)
2015-08-20 05:01:00,297 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:00,297 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:00,300 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:00,302 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:00,359 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1011, Disks:2, vCores:2)
2015-08-20 05:01:00,362 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:00,362 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:01,365 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:01,367 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:01,405 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1012, Disks:2, vCores:2)
2015-08-20 05:01:01,407 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:01,408 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:01,409 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:01,409 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:01,443 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:1012, Disks:2, vCores:2)
2015-08-20 05:01:01,447 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:01,447 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:01,449 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:01,449 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.199 sec - in org.apache.tajo.worker.TestNodeStatusUpdater
Running org.apache.tajo.worker.TestDeletionService
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.072 sec - in org.apache.tajo.worker.TestDeletionService
Running org.apache.tajo.worker.TestNodeResourceManager
2015-08-20 05:01:03,558 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:03,559 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-20 05:01:03,560 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:03,560 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:03,729 INFO: org.apache.tajo.worker.TaskManager (handle(182)) - Running ExecutionBlocks: 1, running tasks:0, availableResource: (Memory:1536, Disks:4, vCores:3)
2015-08-20 05:01:05,621 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:05,623 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:05,623 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:05,623 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-20 05:01:05,664 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:05,665 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-20 05:01:05,665 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:05,665 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:05,667 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:05,669 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:05,669 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:05,669 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-20 05:01:05,702 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:05,704 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-20 05:01:05,704 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:05,705 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:05,706 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:05,706 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:05,708 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:05,709 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-20 05:01:05,741 INFO: org.apache.tajo.worker.NodeResourceManager (serviceInit(66)) - Initialized NodeResourceManager for (Memory:2048, Disks:4, vCores:4)
2015-08-20 05:01:05,742 INFO: org.apache.tajo.worker.TaskExecutor (serviceStart(96)) - Started TaskExecutor[4], Fetcher executor[2]
2015-08-20 05:01:05,743 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(97)) - Queueing threshold:1
2015-08-20 05:01:05,743 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStart(101)) - NodeStatusUpdater started.
2015-08-20 05:01:05,744 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE.reporters in tajo-metrics.properties
2015-08-20 05:01:05,746 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property NODE-JVM.reporters in tajo-metrics.properties
2015-08-20 05:01:05,746 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:05,746 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.223 sec - in org.apache.tajo.worker.TestNodeResourceManager
2015-08-20 05:01:05,779 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-20 05:01:05,782 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-20 05:01:05,782 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-20 05:01:05,784 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
ader initialized will read a total of 2 records.
Aug 20, 2015 4:43:43 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:43:43 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 2
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:02 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 3 ms. row count = 1
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 3 ms. row count = 1
Aug 20, 2015 4:44:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 4 ms. row count = 1
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 20, 2015 4:44:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 20, 2015 4:44:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 212
Aug 20, 2015 4:44:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 5 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, PLAIN_DICTIONARY, RLE], dic { 3 entries, 12B raw, 3B comp}
Aug 20, 2015 4:44:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate_function] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 20, 2015 4:44:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 20, 2015 4:44:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 20, 2015 4:44:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 5 records.
Aug 20, 2015 4:44:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 20, 2015 4:44:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 5
2015-08-20 05:01:05,785 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-20 05:01:05,786 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-20 05:01:05,792 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-20 05:01:05,792 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
2015-08-20 05:01:05,792 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:14893) shutdown
2015-08-20 05:01:05,796 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-20 05:01:05,796 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf903.gq1.ygridcore.net_14894 stopped.
2015-08-20 05:01:05,796 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-20 05:01:05,798 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:14890) shutdown
2015-08-20 05:01:05,798 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:14890) shutdown
2015-08-20 05:01:05,805 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_14892 stopped.
2015-08-20 05:01:05,809 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:54380 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-91668bb2-b325-468f-b05a-faa730ba5ce3:NORMAL:127.0.0.1:54380|RBW]]} size 509
2015-08-20 05:01:05,811 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-20 05:01:05,812 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:14892) shutdown
2015-08-20 05:01:05,813 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session cc0f04e0-2d91-48b0-b631-cb5a1fee8903 is removed.
2015-08-20 05:01:05,813 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 29a024d3-a1a3-46fa-bedc-f96c05c93ba1 is removed.
2015-08-20 05:01:05,814 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:14891) shutdown
2015-08-20 05:01:05,821 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:14889) shutdown
2015-08-20 05:01:05,822 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-20 05:01:05,858 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-20 05:01:05,858 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-20 05:01:05,858 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:14896) shutdown
2015-08-20 05:01:05,859 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-20 05:01:05,860 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-20 05:01:05,864 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-20 05:01:05,864 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-20 05:01:05,866 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:14895) shutdown
2015-08-20 05:01:05,866 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-20 05:01:05,867 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:14894) shutdown
2015-08-20 05:01:05,867 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-20 05:01:05,871 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
Results :
Failed tests:
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestTajoClientV2.testExecuteQueryAsyncWithListener:191 null
Tests run: 1698, Failures: 4, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.989 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.783 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 3.097 s]
[INFO] Tajo Common ....................................... SUCCESS [ 24.221 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.586 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.674 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.725 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.383 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [01:25 min]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.454 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.590 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.497 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [01:02 min]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.196 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.746 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.607 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 2.777 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 2.086 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 4.007 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.420 s]
[INFO] Tajo Core ......................................... SUCCESS [ 10.228 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.928 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.233 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.114 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.013 s]
[INFO] Tajo Client Example ............................... SUCCESS [ 1.005 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 4.828 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.950 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.551 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.624 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:42 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 26:12 min
[INFO] Finished at: 2015-08-20T05:01:07+00:00
[INFO] Final Memory: 139M/1832M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58881863 bytes
Compression is 0.0%
Took 17 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #110
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/110/changes>
Changes:
[jhkim] TAJO-1777: JsonLineDeserializer returns invalid unicode text, if contains control character
------------------------------------------
[...truncated 730790 lines...]
2015-08-19 03:44:44,642 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create QueryCoordinatorProtocol-87 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:44,643 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (QueryCoordinatorProtocol) listens on /127.0.0.1:44889
2015-08-19 03:44:44,643 INFO: org.apache.tajo.master.QueryCoordinatorService (serviceStart(72)) - Instantiated TajoMasterService at localhost/127.0.0.1:44889
2015-08-19 03:44:44,643 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create Tajo-REST-88 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:44,644 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (Tajo-REST) listens on /0:0:0:0:0:0:0:0:60179
2015-08-19 03:44:44,645 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStart(118)) - Tajo Rest Service started.
2015-08-19 03:44:44,646 INFO: BlockStateChange (addToInvalidates(1106)) - BLOCK* addToInvalidates: blk_1073741825_1001 127.0.0.1:50743
2015-08-19 03:44:44,648 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property MASTER.reporters in tajo-metrics.properties
2015-08-19 03:44:44,648 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property MASTER-JVM.reporters in tajo-metrics.properties
2015-08-19 03:44:44,663 INFO: org.apache.tajo.ha.HdfsServiceTracker (initSystemDirectory(93)) - System HA dir 'hdfs://localhost:55595/tajo/system/ha' is created
2015-08-19 03:44:44,664 INFO: org.apache.tajo.ha.HdfsServiceTracker (initSystemDirectory(99)) - System HA Active dir 'hdfs://localhost:55595/tajo/system/ha/active' is created
2015-08-19 03:44:44,665 INFO: org.apache.tajo.ha.HdfsServiceTracker (initSystemDirectory(105)) - System HA Backup dir 'hdfs://localhost:55595/tajo/system/ha/backup' is created
2015-08-19 03:44:44,677 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:50743 is added to blk_1073748705_7881{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3bd29a71-3922-4085-b27d-ddba6e4f95b0:NORMAL:127.0.0.1:50743|RBW]]} size 81
2015-08-19 03:44:44,688 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:50743 is added to blk_1073748706_7882{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-bd724917-fe89-4ee0-8c42-e83442f90b95:NORMAL:127.0.0.1:50743|FINALIZED]]} size 0
2015-08-19 03:44:44,690 INFO: org.apache.tajo.ha.HdfsServiceTracker (register(158)) - This is added to active master (127.0.0.1:44889)
2015-08-19 03:44:44,691 INFO: org.apache.tajo.util.history.HistoryWriter (run(224)) - HistoryWriter_127.0.0.1_44889 started.
2015-08-19 03:44:44,692 INFO: org.apache.tajo.util.history.HistoryCleaner (run(62)) - History cleaner started: expiry day=7
2015-08-19 03:44:44,692 INFO: org.apache.tajo.master.TajoMaster (checkAndInitializeSystemDirectories(254)) - Tajo Root Directory: hdfs://localhost:55595/tajo
2015-08-19 03:44:44,692 INFO: org.apache.tajo.master.TajoMaster (checkAndInitializeSystemDirectories(259)) - FileSystem (hdfs://localhost:55595) is initialized.
2015-08-19 03:44:44,694 INFO: org.apache.tajo.master.TajoMaster (checkAndInitializeSystemDirectories(279)) - Tajo Warehouse dir: hdfs://localhost:55595/tajo/warehouse
2015-08-19 03:44:44,694 INFO: org.apache.tajo.master.TajoMaster (checkAndInitializeSystemDirectories(288)) - Staging dir: hdfs://localhost:55595/tajo/warehouse
2015-08-19 03:44:45,228 INFO: org.apache.tajo.master.rm.TajoResourceManager (serviceInit(97)) - Loaded resource scheduler : class org.apache.tajo.master.scheduler.SimpleScheduler
2015-08-19 03:44:45,229 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create TajoResourceTrackerProtocol-89 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:45,230 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (TajoResourceTrackerProtocol) listens on /127.0.0.1:58867
2015-08-19 03:44:45,230 INFO: org.apache.tajo.master.rm.TajoResourceTracker (serviceInit(99)) - TajoResourceTracker starts up (localhost/127.0.0.1:58867)
2015-08-19 03:44:45,230 INFO: org.apache.tajo.catalog.CatalogServer (serviceInit(125)) - Catalog Store Class: org.apache.tajo.catalog.store.MemStore
Aug 19, 2015 3:44:45 AM org.glassfish.jersey.server.ApplicationHandler initialize
INFO: Initiating Jersey application, version Jersey: 2.6 2014-02-18 21:52:53...
2015-08-19 03:44:45,283 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceInit(109)) - Tajo Rest Service initialized.
2015-08-19 03:44:45,283 INFO: org.apache.tajo.master.TajoMaster (serviceInit(214)) - Tajo Master is initialized.
2015-08-19 03:44:45,283 INFO: org.apache.tajo.master.TajoMaster (serviceStart(316)) - TajoMaster is starting up
2015-08-19 03:44:45,284 INFO: org.apache.tajo.catalog.CatalogServer (createTablespace(211)) - tablespace "default" (hdfs://localhost:55595/tajo/warehouse) is created
2015-08-19 03:44:45,284 INFO: org.apache.tajo.catalog.CatalogServer (createDatabase(402)) - database "default" is created
2015-08-19 03:44:45,292 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create CatalogProtocol-90 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:45,293 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (CatalogProtocol) listens on /127.0.0.1:50212
2015-08-19 03:44:45,293 INFO: org.apache.tajo.catalog.CatalogServer (serviceStart(172)) - Catalog Server startup (127.0.0.1:50212)
2015-08-19 03:44:45,413 INFO: BlockStateChange (invalidateWorkForOneNode(3488)) - BLOCK* BlockManager: ask 127.0.0.1:50743 to delete [blk_1073741825_1001, blk_1073748691_7867]
2015-08-19 03:44:45,590 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create TajoMasterClientProtocol-91 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:45,591 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (TajoMasterClientProtocol) listens on /127.0.0.1:56764
2015-08-19 03:44:45,591 INFO: org.apache.tajo.master.TajoMasterClientService (serviceStart(110)) - Instantiated TajoMasterClientService at localhost/127.0.0.1:56764
2015-08-19 03:44:45,591 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create QueryCoordinatorProtocol-92 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:45,592 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (QueryCoordinatorProtocol) listens on /127.0.0.1:49754
2015-08-19 03:44:45,592 INFO: org.apache.tajo.master.QueryCoordinatorService (serviceStart(72)) - Instantiated TajoMasterService at localhost/127.0.0.1:49754
2015-08-19 03:44:45,592 INFO: org.apache.tajo.rpc.RpcChannelFactory (createServerChannelFactory(141)) - Create Tajo-REST-93 ServerSocketChannelFactory. Worker:2
2015-08-19 03:44:45,593 INFO: org.apache.tajo.rpc.NettyServerBase (start(130)) - Rpc (Tajo-REST) listens on /0:0:0:0:0:0:0:0:43492
2015-08-19 03:44:45,594 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStart(118)) - Tajo Rest Service started.
2015-08-19 03:44:45,595 INFO: BlockStateChange (addToInvalidates(1106)) - BLOCK* addToInvalidates: blk_1073748706_7882 127.0.0.1:50743
2015-08-19 03:44:45,597 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property MASTER.reporters in tajo-metrics.properties
2015-08-19 03:44:45,597 WARN: org.apache.tajo.util.metrics.TajoSystemMetrics (setMetricsReporter(133)) - No property MASTER-JVM.reporters in tajo-metrics.properties
2015-08-19 03:44:45,601 WARN: org.apache.hadoop.hdfs.StateChange (startFileInternal(2602)) - DIR* NameSystem.startFile: /tajo/system/ha/active/active.lock /tajo/system/ha/active/active.lock for client 127.0.0.1 already exists
2015-08-19 03:44:45,609 INFO: org.apache.tajo.ha.HdfsServiceTracker (createLockFile(227)) - Lock file already exists at (hdfs://localhost:55595/tajo/system/ha/active/active.lock)
2015-08-19 03:44:45,622 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:50743 is added to blk_1073748707_7883{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3bd29a71-3922-4085-b27d-ddba6e4f95b0:NORMAL:127.0.0.1:50743|RBW]]} size 81
2015-08-19 03:44:45,624 INFO: org.apache.tajo.ha.HdfsServiceTracker (register(189)) - This is added to backup master (127.0.0.1:49754)
2015-08-19 03:44:45,625 INFO: org.apache.tajo.util.history.HistoryWriter (run(224)) - HistoryWriter_127.0.0.1_49754 started.
2015-08-19 03:44:45,626 INFO: org.apache.tajo.util.history.HistoryCleaner (run(62)) - History cleaner started: expiry day=7
2015-08-19 03:44:45,637 INFO: org.apache.tajo.session.SessionManager (createSession(80)) - Session 953192d2-812a-4700-b681-293199e75b4f is created.
2015-08-19 03:44:45,638 INFO: org.apache.tajo.master.GlobalEngine (executeQuery(182)) - Query: CREATE TABLE default.ha_test1 (age int);
2015-08-19 03:44:45,639 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(280)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-19 03:44:45,639 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - =============================================
2015-08-19 03:44:45,640 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(283)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-19 03:44:45,640 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-19 03:44:45,642 INFO: org.apache.tajo.catalog.CatalogServer (createTable(667)) - relation "default.ha_test1" is added to the catalog (127.0.0.1:37051)
2015-08-19 03:44:45,643 INFO: org.apache.tajo.master.GlobalEngine (executeQuery(182)) - Query: CREATE TABLE default.ha_test2 (age int);
2015-08-19 03:44:45,643 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(280)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-19 03:44:45,643 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - =============================================
2015-08-19 03:44:45,644 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(283)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-19 03:44:45,644 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-19 03:44:45,645 INFO: org.apache.tajo.catalog.CatalogServer (createTable(667)) - relation "default.ha_test2" is added to the catalog (127.0.0.1:37051)
2015-08-19 03:44:45,646 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 953192d2-812a-4700-b681-293199e75b4f is removed.
2015-08-19 03:44:45,647 INFO: org.apache.tajo.session.SessionManager (createSession(80)) - Session 2d5dc1ab-0234-4530-9d0c-42bb71594bb2 is created.
2015-08-19 03:44:45,648 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 2d5dc1ab-0234-4530-9d0c-42bb71594bb2 is removed.
2015-08-19 03:44:45,650 INFO: BlockStateChange (addToInvalidates(1106)) - BLOCK* addToInvalidates: blk_1073748705_7881 127.0.0.1:50743
2015-08-19 03:44:45,652 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:60179) shutdown
2015-08-19 03:44:45,652 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-19 03:44:45,652 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:37051) shutdown
2015-08-19 03:44:45,652 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:37051) shutdown
2015-08-19 03:44:45,654 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-19 03:44:45,654 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:44889) shutdown
2015-08-19 03:44:45,655 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:60161) shutdown
2015-08-19 03:44:45,654 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_44889 stopped.
2015-08-19 03:44:45,657 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:59096) shutdown
2015-08-19 03:44:45,657 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-19 03:44:46,631 INFO: BlockStateChange (addToInvalidates(1106)) - BLOCK* addToInvalidates: blk_1073748707_7883 127.0.0.1:50743
2015-08-19 03:44:46,639 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:50743 is added to blk_1073748708_7884{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-bd724917-fe89-4ee0-8c42-e83442f90b95:NORMAL:127.0.0.1:50743|RBW]]} size 81
2015-08-19 03:44:46,654 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:50743 is added to blk_1073748709_7885{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3bd29a71-3922-4085-b27d-ddba6e4f95b0:NORMAL:127.0.0.1:50743|RBW]]} size 0
2015-08-19 03:44:46,656 INFO: org.apache.tajo.ha.HdfsServiceTracker (register(158)) - This is added to active master (127.0.0.1:49754)
2015-08-19 03:44:46,664 INFO: org.apache.tajo.session.SessionManager (createSession(80)) - Session 066328e7-5f89-47b1-8aaa-b11caac2ccd5 is created.
2015-08-19 03:44:46,666 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 066328e7-5f89-47b1-8aaa-b11caac2ccd5 is removed.
2015-08-19 03:44:46,669 INFO: BlockStateChange (addToInvalidates(1106)) - BLOCK* addToInvalidates: blk_1073748708_7884 127.0.0.1:50743
2015-08-19 03:44:46,671 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:43492) shutdown
2015-08-19 03:44:46,671 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-19 03:44:46,671 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:50212) shutdown
2015-08-19 03:44:46,671 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:50212) shutdown
2015-08-19 03:44:46,672 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_49754 stopped.
2015-08-19 03:44:46,672 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-19 03:44:46,675 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:49754) shutdown
2015-08-19 03:44:46,675 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:56764) shutdown
2015-08-19 03:44:46,676 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:58867) shutdown
2015-08-19 03:44:46,676 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.879 sec - in org.apache.tajo.ha.TestHAServiceHDFSImpl
Running org.apache.tajo.TestTajoIds
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0 sec - in org.apache.tajo.TestTajoIds
2015-08-19 03:44:46,682 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
2015-08-19 03:44:46,683 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-19 03:44:46,686 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-19 03:44:46,686 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-19 03:44:46,686 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
2015-08-19 03:44:46,686 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-19 03:44:46,690 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-19 03:44:46,690 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session e650c61c-5f02-49ec-a97e-87f134c90d64 is removed.
2015-08-19 03:44:46,690 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-19 03:44:46,690 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 3e4b8165-03f9-46c7-a337-da3e5da120a1 is removed.
2015-08-19 03:44:46,692 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf907.gq1.ygridcore.net_47715 stopped.
2015-08-19 03:44:46,692 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-19 03:44:46,692 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:47714) shutdown
2015-08-19 03:44:46,694 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-19 03:44:46,695 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:47711) shutdown
2015-08-19 03:44:46,695 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:47711) shutdown
2015-08-19 03:44:46,699 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_47713 stopped.
2015-08-19 03:44:46,702 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:50743 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-bd724917-fe89-4ee0-8c42-e83442f90b95:NORMAL:127.0.0.1:50743|RBW]]} size 524
2015-08-19 03:44:46,704 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-19 03:44:46,704 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:47713) shutdown
2015-08-19 03:44:46,704 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:47712) shutdown
2015-08-19 03:44:46,708 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:47710) shutdown
2015-08-19 03:44:46,709 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-19 03:44:46,733 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-19 03:44:46,734 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-19 03:44:46,734 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:47717) shutdown
2015-08-19 03:44:46,734 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-19 03:44:46,736 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-19 03:44:46,738 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-19 03:44:46,738 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-19 03:44:46,739 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:47716) shutdown
2015-08-19 03:44:46,739 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-19 03:44:46,739 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:47715) shutdown
2015-08-19 03:44:46,739 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-19 03:44:46,741 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
Results :
Failed tests:
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1691, Failures: 3, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.866 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 2.473 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 4.530 s]
[INFO] Tajo Common ....................................... SUCCESS [ 24.973 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.530 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.912 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.372 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.320 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [01:26 min]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.417 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.649 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 15.068 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 47.283 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 2.644 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.525 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.889 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 2.563 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 2.172 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.849 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.365 s]
[INFO] Tajo Core ......................................... SUCCESS [ 10.139 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.903 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 11.642 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.040 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.968 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 5.071 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.931 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.396 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 3.033 s]
[INFO] Tajo Core Tests ................................... FAILURE [23:36 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 28:02 min
[INFO] Finished at: 2015-08-19T03:44:48+00:00
[INFO] Final Memory: 145M/2238M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58865227 bytes
Compression is 0.0%
Took 16 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #109
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/109/changes>
Changes:
[hyunsik] TAJO-1763: tpch/*.tbl files cannot be founded in maven modules except for core-tests.
[hyunsik] Add missed change log and reindent some logs.
[jhkim] TAJO-1779: Remove "DFSInputStream has been closed already" messages in DelimitedLineReader.
[jihoonson] TAJO-1755: Add documentation for missing built-in functions.
------------------------------------------
[...truncated 723908 lines...]
Aug 18, 2015 5:47:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 18, 2015 5:47:03 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 18, 2015 5:47:04 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 212
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 5 values, 10B raw, 10B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN_DICTIONARY], dic { 3 entries, 12B raw, 3B comp}
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate_function] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [BIT_PACKED, RLE, PLAIN]
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 5 records.
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 18, 2015 5:47:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 5
2015-08-18 06:00:53,651 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-18 06:00:53,653 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-18 06:00:53,653 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
2015-08-18 06:00:53,655 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf909.gq1.ygridcore.net_30203 stopped.
2015-08-18 06:00:53,655 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-18 06:00:53,661 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
2015-08-18 06:00:53,661 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-18 06:00:53,663 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-18 06:00:53,663 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-18 06:00:53,663 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-18 06:00:53,664 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-18 06:00:53,664 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-18 06:00:53,666 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:30202) shutdown
2015-08-18 06:00:53,666 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
2015-08-18 06:00:53,668 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:30199) shutdown
2015-08-18 06:00:53,668 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:30199) shutdown
2015-08-18 06:00:53,669 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:30205) shutdown
2015-08-18 06:00:53,670 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-18 06:00:53,671 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_30201 stopped.
2015-08-18 06:00:53,671 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-18 06:00:53,673 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-18 06:00:53,673 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-18 06:00:53,675 WARN: org.apache.hadoop.hdfs.DFSClient (flushOrSync(2025)) - Unable to persist blocks in hflush for /tajo/system/ha/active/127.0.0.1_39047
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException): No lease on /tajo/system/ha/active/127.0.0.1_39047 (inode 29786): File does not exist. Holder DFSClient_NONMAPREDUCE_-2056499274_1 does not have any open files.
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:3433)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.fsync(FSNamesystem.java:3998)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.fsync(NameNodeRpcServer.java:1210)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.fsync(ClientNamenodeProtocolServerSideTranslatorPB.java:903)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2045)
at org.apache.hadoop.ipc.Client.call(Client.java:1476)
at org.apache.hadoop.ipc.Client.call(Client.java:1407)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
at com.sun.proxy.$Proxy30.fsync(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.fsync(ClientNamenodeProtocolTranslatorPB.java:838)
at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
at com.sun.proxy.$Proxy31.fsync(Unknown Source)
at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:279)
at com.sun.proxy.$Proxy74.fsync(Unknown Source)
at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:279)
at com.sun.proxy.$Proxy74.fsync(Unknown Source)
at org.apache.hadoop.hdfs.DFSOutputStream.flushOrSync(DFSOutputStream.java:2022)
at org.apache.hadoop.hdfs.DFSOutputStream.hsync(DFSOutputStream.java:1898)
at org.apache.hadoop.fs.FSDataOutputStream.hsync(FSDataOutputStream.java:139)
at org.apache.tajo.ha.HdfsServiceTracker.createMasterFile(HdfsServiceTracker.java:244)
at org.apache.tajo.ha.HdfsServiceTracker.register(HdfsServiceTracker.java:155)
at org.apache.tajo.ha.HdfsServiceTracker$PingChecker.run(HdfsServiceTracker.java:374)
at java.lang.Thread.run(Thread.java:745)
2015-08-18 06:00:53,680 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:54368 is added to blk_1073741857_1033{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-4e825a18-4489-4bcd-87b7-5ffefccc8269:NORMAL:127.0.0.1:54368|RBW]]} size 3246361
2015-08-18 06:00:53,680 WARN: org.apache.hadoop.hdfs.DFSClient (flushOrSync(2047)) - Error while syncing
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException): No lease on /tajo/system/ha/active/127.0.0.1_39047 (inode 29786): File does not exist. Holder DFSClient_NONMAPREDUCE_-2056499274_1 does not have any open files.
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:3433)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.fsync(FSNamesystem.java:3998)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.fsync(NameNodeRpcServer.java:1210)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.fsync(ClientNamenodeProtocolServerSideTranslatorPB.java:903)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2045)
at org.apache.hadoop.ipc.Client.call(Client.java:1476)
at org.apache.hadoop.ipc.Client.call(Client.java:1407)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
at com.sun.proxy.$Proxy30.fsync(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.fsync(ClientNamenodeProtocolTranslatorPB.java:838)
at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
at com.sun.proxy.$Proxy31.fsync(Unknown Source)
at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:279)
at com.sun.proxy.$Proxy74.fsync(Unknown Source)
at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:279)
at com.sun.proxy.$Proxy74.fsync(Unknown Source)
at org.apache.hadoop.hdfs.DFSOutputStream.flushOrSync(DFSOutputStream.java:2022)
at org.apache.hadoop.hdfs.DFSOutputStream.hsync(DFSOutputStream.java:1898)
at org.apache.hadoop.fs.FSDataOutputStream.hsync(FSDataOutputStream.java:139)
at org.apache.tajo.ha.HdfsServiceTracker.createMasterFile(HdfsServiceTracker.java:244)
at org.apache.tajo.ha.HdfsServiceTracker.register(HdfsServiceTracker.java:155)
at org.apache.tajo.ha.HdfsServiceTracker$PingChecker.run(HdfsServiceTracker.java:374)
at java.lang.Thread.run(Thread.java:745)
2015-08-18 06:00:53,675 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:30204) shutdown
2015-08-18 06:00:53,681 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-18 06:00:53,681 INFO: BlockStateChange (processAndHandleReportedBlock(3171)) - BLOCK* addBlock: block blk_1073748708_7884 on node 127.0.0.1:54368 size 134217728 does not belong to any file
2015-08-18 06:00:53,682 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:30203) shutdown
2015-08-18 06:00:53,682 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-18 06:00:53,683 INFO: BlockStateChange (add(115)) - BLOCK* InvalidateBlocks: add blk_1073748708_7884 to 127.0.0.1:54368
2015-08-18 06:00:53,685 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
2015-08-18 06:00:53,684 WARN: org.apache.hadoop.hdfs.DFSClient (closeResponder(612)) - Caught exception
java.lang.InterruptedException
at java.lang.Object.wait(Native Method)
at java.lang.Thread.join(Thread.java:1245)
at java.lang.Thread.join(Thread.java:1319)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.closeResponder(DFSOutputStream.java:610)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.closeInternal(DFSOutputStream.java:578)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:574)
2015-08-18 06:00:53,693 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:30200)]: ClosedChannelException: Try to reconnect : /127.0.0.1:30200
2015-08-18 06:00:53,693 ERROR: org.apache.hadoop.hdfs.server.datanode.DataNode (run(278)) - 127.0.0.1:54368:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52082 dst: /127.0.0.1:54368
java.io.IOException: Premature EOF from inputStream
at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:201)
at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:472)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:849)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:804)
at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:137)
at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:74)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:251)
at java.lang.Thread.run(Thread.java:745)
2015-08-18 06:00:53,746 INFO: BlockStateChange (invalidateWorkForOneNode(3488)) - BLOCK* BlockManager: ask 127.0.0.1:54368 to delete [blk_1073748705_7881, blk_1073741825_1001, blk_1073748706_7882, blk_1073748707_7883, blk_1073748708_7884]
2015-08-18 06:00:54,082 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:30201) shutdown
2015-08-18 06:00:54,082 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-18 06:00:54,082 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:30200) shutdown
2015-08-18 06:00:54,089 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:30198) shutdown
2015-08-18 06:00:54,089 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
2015-08-18 06:00:54,694 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:30200)]: ClosedChannelException: Try to reconnect : /127.0.0.1:30200
2015-08-18 06:00:55,699 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:30200)]: ConnectException: Connection refused: /127.0.0.1:30200 Try to reconnect : /127.0.0.1:30200
2015-08-18 06:00:56,702 WARN: org.apache.tajo.rpc.NettyClientBase (doReconnect(198)) - Exception [org.apache.tajo.ipc.TajoMasterClientProtocol(/127.0.0.1:30200)]: ConnectException: Connection refused: /127.0.0.1:30200 Try to reconnect : /127.0.0.1:30200
Results :
Failed tests:
TestTajoClientV2.testExecuteQueryAsyncWithListener:191 null
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests in error:
TestHAServiceHDFSImpl.testAutoFailOver:82->verifyDataBaseAndTable:152 ยป TajoInternal
Tests run: 1691, Failures: 4, Errors: 1, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 2.819 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 2.507 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 3.717 s]
[INFO] Tajo Common ....................................... SUCCESS [ 24.543 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.543 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.550 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 7.114 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.353 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [01:26 min]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.317 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [01:16 min]
[INFO] Tajo Storage Common ............................... SUCCESS [ 8.385 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 56.144 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 1.089 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.487 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 2.291 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 2.857 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.816 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 4.264 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.513 s]
[INFO] Tajo Core ......................................... SUCCESS [ 8.860 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.927 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 26.611 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.087 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 1.002 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.959 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.980 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.303 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.405 s]
[INFO] Tajo Core Tests ................................... FAILURE [21:45 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 27:31 min
[INFO] Finished at: 2015-08-18T06:00:58+00:00
[INFO] Final Memory: 139M/1890M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58864379 bytes
Compression is 0.0%
Took 19 sec
Recording test results
Build failed in Jenkins: Tajo-master-jdk8-nightly #108
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Tajo-master-jdk8-nightly/108/changes>
Changes:
[hyunsik] Bump up to 0.12.0-SNAPSHOT.
[hyunsik] Add missed SNAPSHOT suffix.
------------------------------------------
[...truncated 733298 lines...]
2015-08-17 02:26:03,414 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-17 02:26:03,414 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-17 02:26:03,418 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:35930 is added to blk_1073748707_7883{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-e6518683-d5b2-480e-b961-ce5c42e5115a:NORMAL:127.0.0.1:35930|RBW]]} size 0
2015-08-17 02:26:03,418 INFO: org.apache.tajo.worker.TaskAttemptContext (setState(142)) - Query status of ta_1439777128104_2052_000001_000000_00 is changed to TA_SUCCEEDED
2015-08-17 02:26:03,418 INFO: org.apache.tajo.worker.TaskImpl (run(460)) - ta_1439777128104_2052_000001_000000_00 completed. Worker's task counter - total:1, succeeded: 1, killed: 0, failed: 0
2015-08-17 02:26:03,419 INFO: org.apache.tajo.querymaster.Stage (transition(1298)) - Stage - eb_1439777128104_2052_000001 finalize NONE_SHUFFLE (total=1, success=1, killed=0)
2015-08-17 02:26:03,419 INFO: org.apache.tajo.querymaster.DefaultTaskScheduler (stop(160)) - Task Scheduler stopped
2015-08-17 02:26:03,419 INFO: org.apache.tajo.querymaster.DefaultTaskScheduler (run(122)) - TaskScheduler schedulingThread stopped
2015-08-17 02:26:03,419 INFO: org.apache.tajo.querymaster.Stage (transition(1356)) - Stage completed - eb_1439777128104_2052_000001 (total=1, success=1, killed=0)
2015-08-17 02:26:03,420 INFO: org.apache.tajo.querymaster.Query (handle(855)) - Processing q_1439777128104_2052 of type STAGE_COMPLETED
2015-08-17 02:26:03,420 INFO: org.apache.tajo.engine.planner.global.ParallelExecutionQueue (next(95)) - Next executable block eb_1439777128104_2052_000002
2015-08-17 02:26:03,420 INFO: org.apache.tajo.worker.TaskManager (stopExecutionBlock(161)) - Stopped execution block:eb_1439777128104_2052_000001
2015-08-17 02:26:03,420 INFO: org.apache.tajo.querymaster.Query (transition(802)) - Complete Stage[eb_1439777128104_2052_000001], State: SUCCEEDED, 1/1.
2015-08-17 02:26:03,420 INFO: org.apache.tajo.querymaster.Query (handle(855)) - Processing q_1439777128104_2052 of type QUERY_COMPLETED
2015-08-17 02:26:03,421 INFO: org.apache.tajo.querymaster.Query (finalizeQuery(528)) - Can't find partitions for adding.
2015-08-17 02:26:03,421 INFO: org.apache.tajo.querymaster.Query (handle(873)) - q_1439777128104_2052 Query Transitioned from QUERY_RUNNING to QUERY_SUCCEEDED
2015-08-17 02:26:03,421 INFO: org.apache.tajo.querymaster.QueryMasterTask (handle(295)) - Query completion notified from q_1439777128104_2052 final state: QUERY_SUCCEEDED
2015-08-17 02:26:03,421 INFO: org.apache.tajo.master.QueryInProgress (heartbeat(252)) - Received QueryMaster heartbeat:q_1439777128104_2052,state=QUERY_SUCCEEDED,progress=1.0, queryMaster=asf900.gq1.ygridcore.net
2015-08-17 02:26:03,422 INFO: org.apache.tajo.master.QueryManager (stopQuery(275)) - Stop QueryInProgress:q_1439777128104_2052
2015-08-17 02:26:03,422 INFO: org.apache.tajo.master.QueryInProgress (stopProgress(117)) - =========================================================
2015-08-17 02:26:03,422 INFO: org.apache.tajo.master.QueryInProgress (stopProgress(118)) - Stop query:q_1439777128104_2052
2015-08-17 02:26:03,422 INFO: org.apache.tajo.querymaster.QueryMasterTask (serviceStop(172)) - Stopping QueryMasterTask:q_1439777128104_2052
2015-08-17 02:26:03,422 INFO: org.apache.tajo.querymaster.QueryMasterTask (cleanupQuery(471)) - Cleanup resources of all workers. Query: q_1439777128104_2052, workers: 1
2015-08-17 02:26:03,422 INFO: org.apache.tajo.querymaster.QueryMasterTask (serviceStop(188)) - Stopped QueryMasterTask:q_1439777128104_2052
2015-08-17 02:26:03,640 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-17 02:26:03,640 WARN: org.apache.hadoop.hdfs.DFSClient (close(669)) - DFSInputStream has been closed already
2015-08-17 02:26:03,642 INFO: org.apache.tajo.master.TajoMasterClientService (getQueryResultData(579)) - Send result to client for f70274f8-1fa4-42c2-a854-ed2bf2f561f0,q_1439777128104_2052, 2 rows
2015-08-17 02:26:03,643 INFO: org.apache.tajo.master.TajoMasterClientService (getQueryResultData(579)) - Send result to client for f70274f8-1fa4-42c2-a854-ed2bf2f561f0,q_1439777128104_2052, 0 rows
2015-08-17 02:26:03,644 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session f70274f8-1fa4-42c2-a854-ed2bf2f561f0 is removed.
2015-08-17 02:26:03,645 INFO: org.apache.tajo.master.GlobalEngine (updateQuery(237)) - SQL: DROP TABLE IF EXISTS "TestTajoJdbc".table1
2015-08-17 02:26:03,645 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(280)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-17 02:26:03,646 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - =============================================
2015-08-17 02:26:03,646 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(283)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-17 02:26:03,646 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-17 02:26:03,646 INFO: org.apache.tajo.master.exec.DDLExecutor (dropTable(310)) - relation "TestTajoJdbc.table1" is already exists.
2015-08-17 02:26:03,647 INFO: org.apache.tajo.master.GlobalEngine (updateQuery(237)) - SQL: DROP TABLE IF EXISTS testaltertablepartition
2015-08-17 02:26:03,647 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(280)) - Non Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-17 02:26:03,647 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(282)) - =============================================
2015-08-17 02:26:03,647 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(283)) - Optimized Query:
-----------------------------
Query Block Graph
-----------------------------
|-#ROOT
-----------------------------
Optimization Log:
-----------------------------
2015-08-17 02:26:03,647 INFO: org.apache.tajo.master.GlobalEngine (createLogicalPlan(284)) - =============================================
2015-08-17 02:26:03,648 INFO: org.apache.tajo.catalog.CatalogServer (dropTable(697)) - relation "TestTajoJdbc.testaltertablepartition" is deleted from the catalog (127.0.0.1:38300)
2015-08-17 02:26:03,648 INFO: org.apache.tajo.master.exec.DDLExecutor (dropTable(327)) - relation "TestTajoJdbc.testaltertablepartition" is dropped.
2015-08-17 02:26:03,649 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 5fa8f626-1eff-4560-a6fe-1d0220084375 is removed.
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.542 sec - in org.apache.tajo.jdbc.TestTajoJdbc
2015-08-17 02:26:03,655 INFO: org.apache.tajo.worker.TajoWorker (run(567)) - ============================================
2015-08-17 02:26:03,657 INFO: org.apache.tajo.worker.TajoWorker (run(568)) - TajoWorker received SIGINT Signal
2015-08-17 02:26:03,658 INFO: org.apache.tajo.worker.TajoWorker (run(569)) - ============================================
2015-08-17 02:26:03,658 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 8d61af39-989b-4be2-9b19-9eb04da06e25 is removed.
2015-08-17 02:26:03,659 INFO: org.apache.tajo.master.TajoMaster (run(538)) - ============================================
2015-08-17 02:26:03,661 INFO: org.mortbay.log (info(67)) - Shutdown hook executing
2015-08-17 02:26:03,666 INFO: org.apache.tajo.master.TajoMaster (run(539)) - TajoMaster received SIGINT Signal
2015-08-17 02:26:03,668 INFO: org.mortbay.log (info(67)) - Shutdown hook complete
2015-08-17 02:26:03,668 INFO: org.apache.tajo.master.TajoMaster (run(540)) - ============================================
2015-08-17 02:26:03,670 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-17 02:26:03,670 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_asf900.gq1.ygridcore.net_38304 stopped.
2015-08-17 02:26:03,676 INFO: org.apache.tajo.worker.NodeStatusUpdater (run(262)) - Heartbeat Thread stopped.
2015-08-17 02:26:03,676 INFO: org.apache.tajo.worker.NodeStatusUpdater (serviceStop(111)) - NodeStatusUpdater stopped.
2015-08-17 02:26:03,674 INFO: org.apache.tajo.session.SessionManager (removeSession(86)) - Session 6651846e-8724-49c5-af64-53497fd5ca4b is removed.
2015-08-17 02:26:03,678 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (Tajo-REST) listened on 0:0:0:0:0:0:0:0:38303) shutdown
2015-08-17 02:26:03,680 INFO: org.apache.tajo.ws.rs.TajoRestService (serviceStop(129)) - Tajo Rest Service stopped.
ader initialized will read a total of 2 records.
Aug 17, 2015 2:08:52 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:08:52 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 2
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:07 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 26
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 65B for [l_shipdate_function] BINARY: 1 values, 20B raw, 20B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 1 records.
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:09:08 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 1
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore to file. allocated memory: 212
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [l_orderkey] INT32: 5 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN_DICTIONARY, RLE, BIT_PACKED], dic { 3 entries, 12B raw, 3B comp}
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 123B for [l_shipdate_function] BINARY: 5 values, 76B raw, 76B comp, 1 pages, encodings: [PLAIN, RLE, BIT_PACKED]
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: reading another 1 footers
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: Initiating action with parallelism: 5
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized will read a total of 5 records.
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next block
Aug 17, 2015 2:09:11 AM INFO: org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row count = 5
2015-08-17 02:26:03,678 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterProtocol) listened on 0:0:0:0:0:0:0:0:38306) shutdown
2015-08-17 02:26:03,681 INFO: org.apache.tajo.querymaster.QueryMasterManagerService (serviceStop(106)) - QueryMasterManagerService stopped
2015-08-17 02:26:03,680 INFO: org.apache.tajo.catalog.CatalogServer (serviceStop(178)) - Catalog Server (127.0.0.1:38300) shutdown
2015-08-17 02:26:03,682 INFO: org.apache.tajo.querymaster.QueryMaster (run(417)) - QueryMaster heartbeat thread stopped
2015-08-17 02:26:03,682 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (CatalogProtocol) listened on 127.0.0.1:38300) shutdown
2015-08-17 02:26:03,686 INFO: org.apache.tajo.util.history.HistoryWriter (run(268)) - HistoryWriter_127.0.0.1_38302 stopped.
2015-08-17 02:26:03,686 INFO: org.apache.tajo.querymaster.QueryMaster (serviceStop(168)) - QueryMaster stopped
2015-08-17 02:26:03,686 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(99)) - TajoWorkerClientService stopping
2015-08-17 02:26:03,689 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryMasterClientProtocol) listened on 0:0:0:0:0:0:0:0:38305) shutdown
2015-08-17 02:26:03,689 INFO: org.apache.tajo.worker.TajoWorkerClientService (stop(103)) - TajoWorkerClientService stopped
2015-08-17 02:26:03,689 INFO: BlockStateChange (logAddStoredBlock(2624)) - BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:35930 is added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-c2eba7c8-cf74-484e-b92d-8239c4ce64c3:NORMAL:127.0.0.1:35930|RBW]]} size 704
2015-08-17 02:26:03,691 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoWorkerProtocol) listened on 0:0:0:0:0:0:0:0:38304) shutdown
2015-08-17 02:26:03,691 INFO: org.apache.tajo.worker.TajoWorkerManagerService (serviceStop(93)) - TajoWorkerManagerService stopped
2015-08-17 02:26:03,693 INFO: org.apache.tajo.worker.TajoWorker (serviceStop(377)) - TajoWorker main thread exiting
2015-08-17 02:26:03,693 INFO: org.apache.tajo.util.history.HistoryCleaner (run(136)) - History cleaner stopped
2015-08-17 02:26:03,693 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (QueryCoordinatorProtocol) listened on 127.0.0.1:38302) shutdown
2015-08-17 02:26:03,695 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoMasterClientProtocol) listened on 127.0.0.1:38301) shutdown
2015-08-17 02:26:03,700 INFO: org.apache.tajo.rpc.NettyServerBase (shutdown(173)) - Rpc (TajoResourceTrackerProtocol) listened on 127.0.0.1:38299) shutdown
2015-08-17 02:26:03,700 INFO: org.apache.tajo.master.TajoMaster (serviceStop(406)) - Tajo Master main thread exiting
Results :
Failed tests:
TestDDLBuilder.testBuildDDLForExternalTable:64 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
TestDDLBuilder.testBuildDDLForBaseTable:103 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]');> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]');>
TestDDLBuilder.testBuildDDLQuotedTableName:90 expected:<...) USING TEXT WITH ('[compression.codec'='org.apache.hadoop.io.compress.GzipCodec', 'text.delimiter'='|]') PARTITION BY COLU...> but was:<...) USING TEXT WITH ('[text.delimiter'='|', 'compression.codec'='org.apache.hadoop.io.compress.GzipCodec]') PARTITION BY COLU...>
Tests run: 1691, Failures: 3, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Tajo Main ......................................... SUCCESS [ 1.472 s]
[INFO] Tajo Project POM .................................. SUCCESS [ 1.202 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [ 2.424 s]
[INFO] Tajo Common ....................................... SUCCESS [ 22.096 s]
[INFO] Tajo Algebra ...................................... SUCCESS [ 2.117 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [ 4.632 s]
[INFO] Tajo Plan ......................................... SUCCESS [ 6.055 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [ 1.191 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [01:26 min]
[INFO] Tajo Catalog Client ............................... SUCCESS [ 1.303 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 9.304 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 7.944 s]
[INFO] Tajo HDFS Storage ................................. SUCCESS [ 43.761 s]
[INFO] Tajo PullServer ................................... SUCCESS [ 0.799 s]
[INFO] Tajo Client ....................................... SUCCESS [ 2.513 s]
[INFO] Tajo CLI tools .................................... SUCCESS [ 1.644 s]
[INFO] Tajo JDBC Driver .................................. SUCCESS [ 2.428 s]
[INFO] ASM (thirdparty) .................................. SUCCESS [ 1.827 s]
[INFO] Tajo RESTful Container ............................ SUCCESS [ 3.436 s]
[INFO] Tajo Metrics ...................................... SUCCESS [ 1.333 s]
[INFO] Tajo Core ......................................... SUCCESS [ 8.131 s]
[INFO] Tajo RPC .......................................... SUCCESS [ 0.907 s]
[INFO] Tajo Catalog Drivers Hive ......................... SUCCESS [ 10.215 s]
[INFO] Tajo Catalog Drivers .............................. SUCCESS [ 0.044 s]
[INFO] Tajo Catalog ...................................... SUCCESS [ 0.960 s]
[INFO] Tajo HBase Storage ................................ SUCCESS [ 3.323 s]
[INFO] Tajo Storage ...................................... SUCCESS [ 0.930 s]
[INFO] Tajo Distribution ................................. SUCCESS [ 5.239 s]
[INFO] Tajo Cluster Tests ................................ SUCCESS [ 2.194 s]
[INFO] Tajo Core Tests ................................... FAILURE [20:56 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 24:52 min
[INFO] Finished at: 2015-08-17T02:26:05+00:00
[INFO] Final Memory: 142M/2008M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project tajo-core-tests: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Tajo-master-jdk8-nightly/ws/tajo-core-tests/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :tajo-core-tests
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Tajo-master-jdk8-nightly #82
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 58856592 bytes
Compression is 0.0%
Took 21 sec
Recording test results