You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/12/09 22:13:35 UTC
Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Dataflow
#4485
See <https://builds.apache.org/job/beam_PostCommit_Java_Nexmark_Dataflow/4485/display/redirect?page=changes>
Changes:
[pabloem] [BEAM-8335] Adds support for multi-output TestStream (#9953)
------------------------------------------
[...truncated 16.38 KB...]
> Task :runners:google-cloud-dataflow-java:worker:windmill:processResources
> Task :runners:google-cloud-dataflow-java:worker:windmill:classes
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :model:pipeline:compileJava FROM-CACHE
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :runners:google-cloud-dataflow-java:worker:windmill:shadowJar
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file <https://builds.apache.org/job/beam_PostCommit_Java_Nexmark_Dataflow/ws/src/sdks/java/extensions/sql/build/generated/fmpp/javacc/Parser.jj> . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD is more than 1. Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist. Will create one.
File "ParseException.java" does not exist. Will create one.
File "Token.java" does not exist. Will create one.
File "SimpleCharStream.java" does not exist. Will create one.
Parser generated with 0 errors and 1 warnings.
> Task :sdks:java:extensions:sql:processResources
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:jar
> Task :sdks:java:extensions:join-library:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:core:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :runners:core-java:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:google-cloud-dataflow-java:compileJava FROM-CACHE
> Task :runners:google-cloud-dataflow-java:classes
> Task :runners:google-cloud-dataflow-java:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:google-cloud-dataflow-java:worker:legacy-worker:compileJava FROM-CACHE
> Task :runners:google-cloud-dataflow-java:worker:legacy-worker:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar
> Task :sdks:java:extensions:sql:compileJava FROM-CACHE
> Task :sdks:java:extensions:sql:classes
> Task :sdks:java:extensions:sql:jar
> Task :sdks:java:testing:nexmark:compileJava FROM-CACHE
> Task :sdks:java:testing:nexmark:classes
> Task :sdks:java:testing:nexmark:jar
> Task :runners:google-cloud-dataflow-java:worker:legacy-worker:shadowJar
> Task :sdks:java:testing:nexmark:run
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
2019-12-09T22:13:17.215Z Running query:PASSTHROUGH; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.216Z Running query:AVERAGE_PRICE_FOR_CATEGORY; exportSummaryToBigQuery:true; numEvents:1000000
2019-12-09T22:13:17.216Z Running query:LOCAL_ITEM_SUGGESTION; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.216Z Running query:SELECTION; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.216Z Running query:AVERAGE_SELLING_PRICE_BY_SELLER; exportSummaryToBigQuery:true; numEvents:1000000
2019-12-09T22:13:17.216Z Running query:CURRENCY_CONVERSION; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.216Z Running query:HOT_ITEMS; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.215Z Running query:USER_SESSIONS; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.216Z Running query:HIGHEST_BID; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.215Z Running query:MONITOR_NEW_USERS; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.215Z Running query:WINNING_BIDS; exportSummaryToBigQuery:true; numEvents:1000000
2019-12-09T22:13:17.215Z Running query:LOG_TO_SHARDED_FILES; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.215Z Running query:PROCESSING_TIME_WINDOWS; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.215Z Running query:SESSION_SIDE_INPUT_JOIN; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:17.215Z Running query:BOUNDED_SIDE_INPUT_JOIN; exportSummaryToBigQuery:true; numEvents:10000000
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 1000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 1000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.385Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.389Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.389Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.472Z Generating 1000000 events in batch mode
2019-12-09T22:13:18.481Z Generating 10000000 events in batch mode
2019-12-09T22:13:18.699Z Expected auction duration is 16667 ms
2019-12-09T22:13:18.702Z Expected auction duration is 16667 ms
2019-12-09T22:13:18.702Z Expected auction duration is 16667 ms
==========================================================================================
Run started 2019-12-09T22:13:16.963Z and ran for PT13.964S
Default configuration:
Exception in thread "main" java.lang.RuntimeException: java.lang.RuntimeException: Failed to create a workflow job: (854887132b43cfef): The workflow could not be created. Causes: (57b80f6eb74e3cdb): Dataflow quota error for jobs-per-project quota. Project apache-beam-testing is running 302 jobs. Please check the quota usage via GCP Console. If it exceeds the limit, please wait for a workflow to finish or contact Google Cloud Support to request an increase in quota. If it does not, contact Google Cloud Support.
{"debug":true,"query":null,"sourceType":"DIRECT","sinkType":"DEVNULL","exportSummaryToBigQuery":false,"pubSubMode":"COMBINED","sideInputType":"DIRECT","sideInputRowCount":500,"sideInputNumShards":3,"sideInputUrl":null,"sessionGap":{"standardDays":0,"standardHours":0,"standardMinutes":10,"standardSeconds":600,"millis":600000},"numEvents":100000,"numEventGenerators":100,"rateShape":"SINE","firstEventRate":10000,"nextEventRate":10000,"rateUnit":"PER_SECOND","ratePeriodSec":600,"preloadSeconds":0,"streamTimeout":240,"isRateLimited":false,"useWallclockEventTime":false,"avgPersonByteSize":200,"avgAuctionByteSize":500,"avgBidByteSize":100,"hotAuctionRatio":2,"hotSellersRatio":4,"hotBiddersRatio":4,"windowSizeSec":10,"windowPeriodSec":5,"watermarkHoldbackSec":0,"numInFlightAuctions":100,"numActivePeople":1000,"coderStrategy":"HAND","cpuDelayMs":0,"diskBusyBytes":0,"auctionSkip":123,"fanout":5,"maxAuctionsWaitingTime":600,"occasionalDelaySec":3,"probDelayedEvent":0.1,"maxLogEvents":100000,"usePubsubPublishTime":false,"outOfOrderGroupSize":1}
at org.apache.beam.sdk.nexmark.Main.runAll(Main.java:128)
at org.apache.beam.sdk.nexmark.Main.main(Main.java:415)
Configurations:
Conf Description
0000 query:PASSTHROUGH; exportSummaryToBigQuery:true; numEvents:10000000
0001 query:CURRENCY_CONVERSION; exportSummaryToBigQuery:true; numEvents:10000000
0002 query:SELECTION; exportSummaryToBigQuery:true; numEvents:10000000
0003 query:LOCAL_ITEM_SUGGESTION; exportSummaryToBigQuery:true; numEvents:10000000
0004 query:AVERAGE_PRICE_FOR_CATEGORY; exportSummaryToBigQuery:true; numEvents:1000000
0005 query:HOT_ITEMS; exportSummaryToBigQuery:true; numEvents:10000000
0006 query:AVERAGE_SELLING_PRICE_BY_SELLER; exportSummaryToBigQuery:true; numEvents:1000000
0007 query:HIGHEST_BID; exportSummaryToBigQuery:true; numEvents:10000000
0008 query:MONITOR_NEW_USERS; exportSummaryToBigQuery:true; numEvents:10000000
0009 query:WINNING_BIDS; exportSummaryToBigQuery:true; numEvents:1000000
Caused by: java.lang.RuntimeException: Failed to create a workflow job: (854887132b43cfef): The workflow could not be created. Causes: (57b80f6eb74e3cdb): Dataflow quota error for jobs-per-project quota. Project apache-beam-testing is running 302 jobs. Please check the quota usage via GCP Console. If it exceeds the limit, please wait for a workflow to finish or contact Google Cloud Support to request an increase in quota. If it does not, contact Google Cloud Support.
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:974)
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:188)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:315)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:301)
at org.apache.beam.sdk.nexmark.NexmarkLauncher.run(NexmarkLauncher.java:1178)
at org.apache.beam.sdk.nexmark.Main$Run.call(Main.java:90)
at org.apache.beam.sdk.nexmark.Main$Run.call(Main.java:79)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
0010 query:LOG_TO_SHARDED_FILES; exportSummaryToBigQuery:true; numEvents:10000000
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
0011 query:USER_SESSIONS; exportSummaryToBigQuery:true; numEvents:10000000
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
0012 query:PROCESSING_TIME_WINDOWS; exportSummaryToBigQuery:true; numEvents:10000000
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
0013 query:BOUNDED_SIDE_INPUT_JOIN; exportSummaryToBigQuery:true; numEvents:10000000
at java.lang.Thread.run(Thread.java:748)
0014 query:SESSION_SIDE_INPUT_JOIN; exportSummaryToBigQuery:true; numEvents:10000000
Caused by: com.google.api.client.googleapis.json.GoogleJsonResponseException: 400 Bad Request
{
Performance:
"code" : 400,
"errors" : [ {
Conf Runtime(sec) (Baseline) Events(/sec) (Baseline) Results (Baseline)
"domain" : "global",
0000 *** not run ***
0001 *** not run ***
0002 *** not run ***
"message" : "(854887132b43cfef): The workflow could not be created. Causes: (57b80f6eb74e3cdb): Dataflow quota error for jobs-per-project quota. Project apache-beam-testing is running 302 jobs. Please check the quota usage via GCP Console. If it exceeds the limit, please wait for a workflow to finish or contact Google Cloud Support to request an increase in quota. If it does not, contact Google Cloud Support.",
0003 *** not run ***
"reason" : "failedPrecondition"
0004 *** not run ***
} ],
0005 *** not run ***
0006 *** not run ***
0007 *** not run ***
"message" : "(854887132b43cfef): The workflow could not be created. Causes: (57b80f6eb74e3cdb): Dataflow quota error for jobs-per-project quota. Project apache-beam-testing is running 302 jobs. Please check the quota usage via GCP Console. If it exceeds the limit, please wait for a workflow to finish or contact Google Cloud Support to request an increase in quota. If it does not, contact Google Cloud Support.",
0008 *** not run ***
"status" : "FAILED_PRECONDITION"
0009 *** not run ***
}
0010 *** not run ***
at com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:150)
0011 *** not run ***
at com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
0012 *** not run ***
at com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
0013 *** not run ***
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest$1.interceptResponse(AbstractGoogleClientRequest.java:417)
0014 *** not run ***
at com.google.api.client.http.HttpRequest.execute(HttpRequest.java:1132)
==========================================================================================
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:515)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:448)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:565)
at org.apache.beam.runners.dataflow.DataflowClient.createJob(DataflowClient.java:61)
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:960)
... 12 more
> Task :sdks:java:testing:nexmark:run FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':sdks:java:testing:nexmark:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 1m 4s
81 actionable tasks: 55 executed, 26 from cache
Publishing build scan...
https://scans.gradle.com/s/divtpmmnley5s
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Jenkins build is back to normal :
beam_PostCommit_Java_Nexmark_Dataflow #4486
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_Nexmark_Dataflow/4486/display/redirect>
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org