You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/08/20 07:33:59 UTC

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Dataflow #204

See <https://builds.apache.org/job/beam_PostCommit_Java_Nexmark_Dataflow/204/display/redirect?page=changes>

Changes:

[mxm] Correct field visibility in SingletonKeyedWorkitem

[mxm] Remove FlinkRunner dependency of FlinkPipelineTranslator

[mxm] Rename BatchFlinkExecutableStageContext FlinkBatchExecutableStageContext

------------------------------------------
[...truncated 1.46 MB...]
2018-08-20T07:33:54.864Z DONE SqlQuery7

==========================================================================================
Run started 2018-08-20T07:27:07.309Z and ran for PT407.603S

Default configuration:
{"debug":true,"query":0,"sourceType":"DIRECT","sinkType":"DEVNULL","exportSummaryToBigQuery":false,"pubSubMode":"COMBINED","numEvents":100000,"numEventGenerators":100,"rateShape":"SINE","firstEventRate":10000,"nextEventRate":10000,"rateUnit":"PER_SECOND","ratePeriodSec":600,"preloadSeconds":0,"streamTimeout":240,"isRateLimited":false,"useWallclockEventTime":false,"avgPersonByteSize":200,"avgAuctionByteSize":500,"avgBidByteSize":100,"hotAuctionRatio":2,"hotSellersRatio":4,"hotBiddersRatio":4,"windowSizeSec":10,"windowPeriodSec":5,"watermarkHoldbackSec":0,"numInFlightAuctions":100,"numActivePeople":1000,"coderStrategy":"HAND","cpuDelayMs":0,"diskBusyBytes":0,"auctionSkip":123,"fanout":5,"maxAuctionsWaitingTime":600,"occasionalDelaySec":3,"probDelayedEvent":0.1,"maxLogEvents":100000,"usePubsubPublishTime":false,"outOfOrderGroupSize":1}

Configurations:
  Conf  Description
  0000  query:0; exportSummaryToBigQuery:true; numEvents:10000000
  0001  query:1; exportSummaryToBigQuery:true; numEvents:10000000
  0002  query:2; exportSummaryToBigQuery:true; numEvents:10000000
  0003  query:3; exportSummaryToBigQuery:true; numEvents:10000000
  0004  query:4; exportSummaryToBigQuery:true; numEvents:1000000
  0005  query:5; exportSummaryToBigQuery:true; numEvents:10000000
  0006  query:6; exportSummaryToBigQuery:true; numEvents:1000000
  0007  query:7; exportSummaryToBigQuery:true; numEvents:10000000
  0008  query:8; exportSummaryToBigQuery:true; numEvents:10000000
  0009  query:9; exportSummaryToBigQuery:true; numEvents:1000000
  0010  query:10; exportSummaryToBigQuery:true; numEvents:10000000
  0011  query:11; exportSummaryToBigQuery:true; numEvents:10000000
  0012  query:12; exportSummaryToBigQuery:true; numEvents:10000000

Performance:
  Conf  Runtime(sec)    (Baseline)  Events(/sec)    (Baseline)       Results    (Baseline)
  0000          98.3                    101732.5                     9200000              
  0001          -1.0                        -1.0                          -1              
        *** Job was unexpectedly updated ***
  0002          40.3                    248342.3                       83527              
  0003          -1.0                        -1.0                          -1              
        *** Job was unexpectedly updated ***
  0004  *** not run ***
  0005         160.5                     62316.6                        4567              
  0006  *** not run ***
  0007         207.7                     48136.6                         100              
  0008  *** not run ***
  0009  *** not run ***
  0010  *** not run ***
  0011  *** not run ***
  0012  *** not run ***
==========================================================================================

Aug 20, 2018 7:33:54 AM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 127 files. Enable logging at DEBUG level to see which files will be staged.
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 127 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 127 files cached, 0 files newly uploaded
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Create.Values/Read(CreateSource) as step s1
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/PrepareWrite/ParDo(Anonymous) as step s2
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/JobIdCreationRoot/Read(CreateSource) as step s3
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/CreateJobId as step s4
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map as step s5
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey as step s6
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues as step s7
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map as step s8
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey) as step s9
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly as step s10
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow) as step s11
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/CreateDataflowView as step s12
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/Create.Values/Read(CreateSource) as step s13
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/GetTempFilePrefix as step s14
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map as step s15
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey as step s16
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues as step s17
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map as step s18
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey) as step s19
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly as step s20
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow) as step s21
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/CreateDataflowView as step s22
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/rewindowIntoGlobal/Window.Assign as step s23
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/WriteBundlesToFiles as step s24
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/GroupByDestination as step s25
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/WriteGroupedRecords as step s26
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/FlattenFiles as step s27
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyResults/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) as step s28
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyResults/View.AsIterable/CreateDataflowView as step s29
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyResults/Create.Values/Read(CreateSource) as step s30
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyResults/ParDo(Anonymous) as step s31
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/WritePartitionUntriggered as step s32
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsReshuffle/Window.Into()/Window.Assign as step s33
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsReshuffle/GroupByKey as step s34
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsReshuffle/ExpandIterable as step s35
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsWriteTables/ParMultiDo(WriteTables) as step s36
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsWriteTables/WithKeys/AddKeys/Map as step s37
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsWriteTables/Window.Into()/Window.Assign as step s38
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsWriteTables/GroupByKey as step s39
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsWriteTables/Values/Values/Map as step s40
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/MultiPartitionsWriteTables/ParDo(GarbageCollectTemporaryFiles) as step s41
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyRenameInput/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) as step s42
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyRenameInput/View.AsIterable/CreateDataflowView as step s43
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource) as step s44
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/ReifyRenameInput/ParDo(Anonymous) as step s45
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/WriteRenameUntriggered as step s46
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionsReshuffle/Window.Into()/Window.Assign as step s47
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionsReshuffle/GroupByKey as step s48
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionsReshuffle/ExpandIterable as step s49
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionWriteTables/ParMultiDo(WriteTables) as step s50
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionWriteTables/WithKeys/AddKeys/Map as step s51
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionWriteTables/Window.Into()/Window.Assign as step s52
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionWriteTables/GroupByKey as step s53
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionWriteTables/Values/Values/Map as step s54
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/SinglePartitionWriteTables/ParDo(GarbageCollectTemporaryFiles) as step s55
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding savePerfsToBigQuery/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource) as step s56
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding DropInputs as step s57
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging pipeline description to gs://temp-storage-for-perf-tests/nexmark/staging/
Aug 20, 2018 7:33:55 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <175608 bytes, hash 2QI3leXXdrD7x6ytQpmRVQ> to gs://temp-storage-for-perf-tests/nexmark/staging/pipeline-2QI3leXXdrD7x6ytQpmRVQ.pb
Dataflow SDK version: 2.7.0-SNAPSHOT
Aug 20, 2018 7:33:57 AM org.apache.beam.runners.dataflow.DataflowRunner run
Submitted job: 2018-08-20_00_33_56-13146900705683295694
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-20_00_33_56-13146900705683295694?project=apache-beam-testing
Aug 20, 2018 7:33:57 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2018-08-20_00_33_56-13146900705683295694

==========================================================================================
Run started 2018-08-20T07:27:07.309Z and ran for PT409.987S

Default configuration:
{"debug":true,"query":0,"sourceType":"DIRECT","sinkType":"DEVNULL","exportSummaryToBigQuery":false,"pubSubMode":"COMBINED","numEvents":100000,"numEventGenerators":100,"rateShape":"SINE","firstEventRate":10000,"nextEventRate":10000,"rateUnit":"PER_SECOND","ratePeriodSec":600,"preloadSeconds":0,"streamTimeout":240,"isRateLimited":false,"useWallclockEventTime":false,"avgPersonByteSize":200,"avgAuctionByteSize":500,"avgBidByteSize":100,"hotAuctionRatio":2,"hotSellersRatio":4,"hotBiddersRatio":4,"windowSizeSec":10,"windowPeriodSec":5,"watermarkHoldbackSec":0,"numInFlightAuctions":100,"numActivePeople":1000,"coderStrategy":"HAND","cpuDelayMs":0,"diskBusyBytes":0,"auctionSkip":123,"fanout":5,"maxAuctionsWaitingTime":600,"occasionalDelaySec":3,"probDelayedEvent":0.1,"maxLogEvents":100000,"usePubsubPublishTime":false,"outOfOrderGroupSize":1}

Configurations:
  Conf  Description
  0000  query:0; exportSummaryToBigQuery:true; numEvents:10000000
  0001  query:1; exportSummaryToBigQuery:true; numEvents:10000000
  0002  query:2; exportSummaryToBigQuery:true; numEvents:10000000
  0003  query:3; exportSummaryToBigQuery:true; numEvents:10000000
  0004  query:4; exportSummaryToBigQuery:true; numEvents:1000000
  0005  query:5; exportSummaryToBigQuery:true; numEvents:10000000
  0006  query:6; exportSummaryToBigQuery:true; numEvents:1000000
  0007  query:7; exportSummaryToBigQuery:true; numEvents:10000000
  0008  query:8; exportSummaryToBigQuery:true; numEvents:10000000
  0009  query:9; exportSummaryToBigQuery:true; numEvents:1000000
  0010  query:10; exportSummaryToBigQuery:true; numEvents:10000000
  0011  query:11; exportSummaryToBigQuery:true; numEvents:10000000
  0012  query:12; exportSummaryToBigQuery:true; numEvents:10000000

Performance:
  Conf  Runtime(sec)    (Baseline)  Events(/sec)    (Baseline)       Results    (Baseline)
  0000          98.3                    101732.5                     9200000              
  0001          -1.0                        -1.0                          -1              
        *** Job was unexpectedly updated ***
  0002          40.3                    248342.3                       83527              
  0003          -1.0                        -1.0                          -1              
        *** Job was unexpectedly updated ***
  0004  *** not run ***
  0005         160.5                     62316.6                        4567              
  0006  *** not run ***
Exception in thread "main"   0007         207.7                     48136.6                         100              
  0008  *** not run ***
  0009  *** not run ***
  0010  *** not run ***
  0011  *** not run ***
  0012  *** not run ***
==========================================================================================

java.lang.RuntimeException: Execution was not successful
	at org.apache.beam.sdk.nexmark.Main.runAll(Main.java:174)
	at org.apache.beam.sdk.nexmark.Main.main(Main.java:477)

> Task :beam-sdks-java-nexmark:run FAILED
:beam-sdks-java-nexmark:run (Thread[Task worker for ':' Thread 3,5,main]) completed. Took 6 mins 50.503 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':beam-sdks-java-nexmark:run'.
> Process 'command '/usr/local/asfpackages/java/jdk1.8.0_152/bin/java'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 7m 1s
65 actionable tasks: 2 executed, 63 up-to-date

Publishing build scan...
https://gradle.com/s/bdrwxqnd2dm6y

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

Jenkins build is back to normal : beam_PostCommit_Java_Nexmark_Dataflow #205

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_Nexmark_Dataflow/205/display/redirect?page=changes>