You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/02/25 08:43:54 UTC

Build failed in Jenkins: beam_LoadTests_Go_CoGBK_Flink_batch #80

See <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/80/display/redirect?page=changes>

Changes:

[relax] First step in Vortex sink - refactor CreateTablea and add TableRow ->

[tysonjh] [BEAM-11377] Fix retry & cleanup issues.

[relax] address comments

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/azure/

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/influxdb

[Kenneth Knowles] Add test for side input created once but consumed twice

[shehzaad] [BEAM-10961] (1) fix azure-storage-common version (2) add reference to

[shehzaad] [BEAM-10961] nit: spacing

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/mqtt

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/solr

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/splunk

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/tika

[shehzaad] [BEAM-10961] enable strict dependency checking for sdks/java/io/xml

[shehzaad] [BEAM-10961] fix spacing

[nir.gzt] [BEAM-11859] Fixed bug in python S3 IO

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/hcatalog

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/kafka

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/jms

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/jdbc

[noreply] [BEAM-10961] enable strict dependency checking for

[noreply] [BEAM-10961] enable strict dependency checking for

[noreply] [BEAM-10961] enable strict dependency checking for

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/kinesis

[noreply] [BEAM-10961] enable strict dependency checking for

[noreply] [BEAM-10961] enable strict dependency checking for

[noreply] Returning successful writes in FhirIO.Write.Result (#14034)

[noreply] Merge pull request #14046 from [BEAM-11791] Adding a microbenchmark for

[noreply] [BEAM-11344] Apply "Become a Committer" changes from Website Revamp

[noreply] [BEAM-10937] Add Tour of Beam page (#13747)

[Kenneth Knowles] Remove metadata-driven triggers from capability matrix, because they do

[Kenneth Knowles] Remove retractions from capability matrix, because they do not exist yet

[Kenneth Knowles] Remove JStorm runner from capability matrix, because it is on a branch

[Kenneth Knowles] Remove MapReduce runner from capability matrix, because it is on a

[Kenneth Knowles] Merge redundant model feature columns in capability matrix

[noreply] Merge pull request #14033 from [BEAM-11408] Integrate Python BigQuery

[Kenneth Knowles] Log a warning when Dataflow returns an unrecognized state

[Kenneth Knowles] Show string from Dataflow service when job terminates in unrecognized

[Brian Hulette] Fix preview


------------------------------------------
[...truncated 55.37 KB...]
  >
  coders: <
    key: "c4"
    value: <
      spec: <
        urn: "beam:coder:length_prefix:v1"
      >
      component_coder_ids: "c3"
    >
  >
  coders: <
    key: "c5"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c4"
      component_coder_ids: "c0"
    >
  >
  coders: <
    key: "c6"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c5"
    >
  >
  coders: <
    key: "c7"
    value: <
      spec: <
        urn: "beam:coder:iterable:v1"
      >
      component_coder_ids: "c5"
    >
  >
  coders: <
    key: "c8"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c7"
    >
  >
  coders: <
    key: "c9"
    value: <
      spec: <
        urn: "beam:go:coder:cogbklist:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c0"
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n>gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_****_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006****"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "s2"
root_transform_ids: "e8"
root_transform_ids: "e4"
root_transform_ids: "e9_cogbk"
root_transform_ids: "e10"
root_transform_ids: "e11"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2021/02/25 08:31:18 Prepared job with id: load-tests-go-flink-batch-cogbk-2-0225065610_9940b085-75e1-4880-9118-1c61e1abfa0b and staging token: load-tests-go-flink-batch-cogbk-2-0225065610_9940b085-75e1-4880-9118-1c61e1abfa0b
2021/02/25 08:31:18 Using specified **** binary: 'linux_amd64/cogbk'
2021/02/25 08:31:21 Staged binary artifact with token: 
2021/02/25 08:31:22 Submitted job: load0tests0go0flink0batch0cogbk0200225065610-root-0225083121-af4df7c4_04606e0d-03eb-4b67-be2f-aae7fbdca39f
2021/02/25 08:31:22 Job state: STOPPED
2021/02/25 08:31:22 Job state: STARTING
2021/02/25 08:31:22 Job state: RUNNING
2021/02/25 08:43:52  (): java.util.concurrent.ExecutionException: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 301614f32741fa99f2750ba9d0f228ba)
	at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
	at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
	at org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:864)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:199)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:118)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:85)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 301614f32741fa99f2750ba9d0f228ba)
	at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$null$6(ClusterClientJobClientAdapter.java:112)
	at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
	at java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
	at org.apache.flink.client.program.rest.RestClusterClient.lambda$pollResourceAsync$21(RestClusterClient.java:565)
	at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
	at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
	at org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:291)
	at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
	at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:575)
	at java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:943)
	at java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456)
	... 3 more
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
	at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$null$6(ClusterClientJobClientAdapter.java:110)
	... 19 more
Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy
	at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:110)
	at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:76)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:192)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:186)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:180)
	at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:496)
	at org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:380)
	at sun.reflect.GeneratedMethodAccessor23.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:284)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:199)
	at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
	at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
	at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
	at akka.actor.ActorCell.invoke(ActorCell.scala:561)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
	at akka.dispatch.Mailbox.run(Mailbox.scala:225)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
	at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.RuntimeException: Emitting the record caused an I/O exception: Failed to serialize element. Serialized size (> 268435834 bytes) exceeds JVM heap space
	at org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:69)
	at org.apache.flink.runtime.operators.util.metrics.CountingCollector.collect(CountingCollector.java:35)
	at org.apache.beam.runners.flink.translation.functions.SortingFlinkCombineRunner.combine(SortingFlinkCombineRunner.java:141)
	at org.apache.beam.runners.flink.translation.functions.FlinkReduceFunction.reduce(FlinkReduceFunction.java:109)
	at org.apache.flink.api.java.operators.translation.PlanUnwrappingReduceGroupOperator$TupleUnwrappingNonCombinableGroupReducer.reduce(PlanUnwrappingReduceGroupOperator.java:111)
	at org.apache.flink.runtime.operators.GroupReduceDriver.run(GroupReduceDriver.java:131)
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:504)
	at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
	at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:708)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:533)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: Failed to serialize element. Serialized size (> 268435834 bytes) exceeds JVM heap space
	at org.apache.flink.core.memory.DataOutputSerializer.resize(DataOutputSerializer.java:334)
	at org.apache.flink.core.memory.DataOutputSerializer.write(DataOutputSerializer.java:152)
	at org.apache.beam.runners.flink.translation.wrappers.DataOutputViewWrapper.write(DataOutputViewWrapper.java:44)
	at java.io.DataOutputStream.write(DataOutputStream.java:107)
	at java.io.ByteArrayOutputStream.writeTo(ByteArrayOutputStream.java:167)
	at org.apache.beam.sdk.coders.LengthPrefixCoder.encode(LengthPrefixCoder.java:58)
	at org.apache.beam.sdk.coders.KvCoder.encode(KvCoder.java:72)
	at org.apache.beam.sdk.coders.KvCoder.encode(KvCoder.java:63)
	at org.apache.beam.sdk.coders.KvCoder.encode(KvCoder.java:37)
	at org.apache.beam.sdk.coders.IterableLikeCoder.encode(IterableLikeCoder.java:114)
	at org.apache.beam.sdk.coders.IterableLikeCoder.encode(IterableLikeCoder.java:60)
	at org.apache.beam.sdk.coders.Coder.encode(Coder.java:136)
	at org.apache.beam.sdk.coders.KvCoder.encode(KvCoder.java:73)
	at org.apache.beam.sdk.coders.KvCoder.encode(KvCoder.java:37)
	at org.apache.beam.sdk.util.WindowedValue$FullWindowedValueCoder.encode(WindowedValue.java:591)
	at org.apache.beam.sdk.util.WindowedValue$FullWindowedValueCoder.encode(WindowedValue.java:582)
	at org.apache.beam.sdk.util.WindowedValue$FullWindowedValueCoder.encode(WindowedValue.java:542)
	at org.apache.beam.runners.flink.translation.types.CoderTypeSerializer.serialize(CoderTypeSerializer.java:111)
	at org.apache.flink.runtime.plugable.SerializationDelegate.write(SerializationDelegate.java:54)
	at org.apache.flink.runtime.io.network.api.serialization.SpanningRecordSerializer.serializeRecord(SpanningRecordSerializer.java:71)
	at org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:113)
	at org.apache.flink.runtime.io.network.api.writer.ChannelSelectorRecordWriter.emit(ChannelSelectorRecordWriter.java:60)
	at org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:65)
	... 10 more
Caused by: java.lang.OutOfMemoryError: Java heap space
2021/02/25 08:43:52  (): java.lang.OutOfMemoryError: Java heap space
2021/02/25 08:43:52 Job state: FAILED
2021/02/25 08:43:52 Failed to execute job: job load0tests0go0flink0batch0cogbk0200225065610-root-0225083121-af4df7c4_04606e0d-03eb-4b67-be2f-aae7fbdca39f failed
panic: Failed to execute job: job load0tests0go0flink0batch0cogbk0200225065610-root-0225083121-af4df7c4_04606e0d-03eb-4b67-be2f-aae7fbdca39f failed

goroutine 1 [running]:
github.com/apache/beam/sdks/go/test/load/vendor/github.com/apache/beam/sdks/go/pkg/beam/log.Fatalf(0x11a2e00, 0xc00003e0b0, 0x1060143, 0x19, 0xc000187ee8, 0x1, 0x1)
	<https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/sdks/go/test/load/.gogradle/project_gopath/src/github.com/apache/beam/sdks/go/test/load/vendor/github.com/apache/beam/sdks/go/pkg/beam/log/log.go>:153 +0xec
main.main()
	<https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/sdks/go/test/load/.gogradle/project_gopath/src/github.com/apache/beam/sdks/go/test/load/cogbk/cogbk.go>:99 +0x577

> Task :sdks:go:test:load:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/sdks/go/test/load/build.gradle'> line: 65

* What went wrong:
Execution failed for task ':sdks:go:test:load:run'.
> Process 'command 'sh'' finished with non-zero exit value 2

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 16m 24s
6 actionable tasks: 6 executed

Publishing build scan...
https://gradle.com/s/6vzpxinqg2klm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_LoadTests_Go_CoGBK_Flink_batch #83

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/83/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Go_CoGBK_Flink_batch #82

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/82/display/redirect?page=changes>

Changes:

[Kenneth Knowles] Remove use of model SYNCHRONIZED_PROCESSING_TIME

[Kenneth Knowles] Remove SYNCHRONIZED_PROCESSING_TIME from model proto

[samuelw] [BEAM-11707] Change WindmillStateCache cache invalidation to be based

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] undo line moves (originally intended for alphabeticization)

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] enable strict dependency checking for

[shehzaad] [BEAM-10961] fix stray reordering of lines

[Robert Bradshaw] Revert "Revert "[BEAM-2914] Add portable merging window support to

[zyichi] Add validate runner test for testing custom merging windows fn without

[zyichi] Fix up! formatting

[Boyuan Zhang] Do not stage dataflow worker jar when use runner_v2.

[Robert Bradshaw] Use the windowing strategy of the input, not output, PCollection of GBK.

[Kenneth Knowles] Recognize JOB_STATE_PENDING from Dataflow and map to RUNNING

[Pablo Estrada] Attempting improvements on DirectRunner Python dash

[shehzaad] [BEAM-10961] add explicit compile for auto_value_annotations in

[shehzaad] [BEAM-10961] add reference to BEAM-11761

[noreply] [BEAM-11531] Use pandas 1.2 for python>=3.7 (#14099)

[noreply] [BEAM-11861] Add methods to explicitly provide coder for ParquetIO's

[noreply] Merge pull request #13802: [BEAM-1474]. Adding MapState and SetState

[noreply] [BEAM-10961] enable strict dependency checking for

[Pablo Estrada] Attempting improvements on DirectRunner Python dash

[Kenneth Knowles] Initial watermark latency benchmark

[noreply] [BEAM-10961] Strict dependency checking for sdks/java/io/gcp (#13791)

[Robert Bradshaw] Improve test, error on ALREADY_MERGED.

[Kenneth Knowles] Fix compile breakage in WindmillStateInternals

[Kenneth Knowles] Fix checkstyle in watermark latency benchmark

[Kenneth Knowles] Remove InvalidWindows from Java SDK, instead track "already merged" bit


------------------------------------------
[...truncated 145.42 KB...]
        payload: "Cgd2YXJpbnR6EgIIAhqKAQp2Z2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvbG9hZC92ZW5kb3IvZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3BrZy9iZWFtL2NvcmUvcnVudGltZS9jb2RlcnguZW5jVmFySW50WhIQCBYiBAgZQA8qBggUEgIICCKWAQp2Z2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvbG9hZC92ZW5kb3IvZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3BrZy9iZWFtL2NvcmUvcnVudGltZS9jb2RlcnguZGVjVmFySW50WhIcCBYiBAgZQAMiBggUEgIICCoECBlADyoECBlAAQ=="
      >
    >
  >
  coders: <
    key: "c4"
    value: <
      spec: <
        urn: "beam:coder:length_prefix:v1"
      >
      component_coder_ids: "c3"
    >
  >
  coders: <
    key: "c5"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c4"
      component_coder_ids: "c0"
    >
  >
  coders: <
    key: "c6"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c5"
    >
  >
  coders: <
    key: "c7"
    value: <
      spec: <
        urn: "beam:coder:iterable:v1"
      >
      component_coder_ids: "c5"
    >
  >
  coders: <
    key: "c8"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c7"
    >
  >
  coders: <
    key: "c9"
    value: <
      spec: <
        urn: "beam:go:coder:cogbklist:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c0"
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n>gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_****_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006****"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "s2"
root_transform_ids: "e4"
root_transform_ids: "e8"
root_transform_ids: "e9_cogbk"
root_transform_ids: "e10"
root_transform_ids: "e11"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2021/02/27 08:52:16 Prepared job with id: load-tests-go-flink-batch-cogbk-4-0227065603_ffcd9f24-9025-4984-9f90-f1778bc10d3a and staging token: load-tests-go-flink-batch-cogbk-4-0227065603_ffcd9f24-9025-4984-9f90-f1778bc10d3a
2021/02/27 08:52:16 Using specified **** binary: 'linux_amd64/cogbk'
2021/02/27 08:52:18 Staged binary artifact with token: 
2021/02/27 08:52:18 Submitted job: load0tests0go0flink0batch0cogbk0400227065603-root-0227085218-62380d27_e82a681f-b30f-4069-9f73-a7dd7e0158a3
2021/02/27 08:52:18 Job state: STOPPED
2021/02/27 08:52:18 Job state: STARTING
2021/02/27 08:52:18 Job state: RUNNING
2021/02/27 09:02:15  (): java.util.concurrent.ExecutionException: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 5f11819a9bd0a81aa51329cd3777e3f5)
	at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
	at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
	at org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:864)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:199)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:118)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:85)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 5f11819a9bd0a81aa51329cd3777e3f5)
	at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$null$6(ClusterClientJobClientAdapter.java:112)
	at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
	at java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
	at org.apache.flink.client.program.rest.RestClusterClient.lambda$pollResourceAsync$21(RestClusterClient.java:565)
	at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
	at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
	at org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:291)
	at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
	at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:575)
	at java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:943)
	at java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456)
	... 3 more
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
	at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$null$6(ClusterClientJobClientAdapter.java:110)
	... 19 more
Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy
	at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:110)
	at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:76)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:192)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:186)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:180)
	at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:496)
	at org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:380)
	at sun.reflect.GeneratedMethodAccessor24.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:284)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:199)
	at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
	at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
	at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
	at akka.actor.ActorCell.invoke(ActorCell.scala:561)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
	at akka.dispatch.Mailbox.run(Mailbox.scala:225)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
	at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.Exception: The data preparation for task 'GroupReduce (GroupReduce at e9_gbk)' , caused an error: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-go-cogbk-flink-batch-82-w-2.c.apache-beam-testing.internal/10.128.0.186:42695'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480)
	at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
	at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:708)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:533)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.RuntimeException: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-go-cogbk-flink-batch-82-w-2.c.apache-beam-testing.internal/10.128.0.186:42695'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger.getIterator(UnilateralSortMerger.java:650)
	at org.apache.flink.runtime.operators.BatchTask.getInput(BatchTask.java:1110)
	at org.apache.flink.runtime.operators.GroupReduceDriver.prepare(GroupReduceDriver.java:99)
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:474)
	... 4 more
Caused by: java.io.IOException: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-go-cogbk-flink-batch-82-w-2.c.apache-beam-testing.internal/10.128.0.186:42695'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger$ThreadBase.run(UnilateralSortMerger.java:831)
Caused by: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connection unexpectedly closed by remote task manager 'beam-loadtests-go-cogbk-flink-batch-82-w-2.c.apache-beam-testing.internal/10.128.0.186:42695'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.channelInactive(CreditBasedPartitionRequestClientHandler.java:136)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:257)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:243)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:236)
	at org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInputClosed(ByteToMessageDecoder.java:393)
	at org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInactive(ByteToMessageDecoder.java:358)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:257)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:243)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:236)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1416)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:257)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:243)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:912)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:816)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:416)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:515)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:918)
	at org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
	at java.lang.Thread.run(Thread.java:748)
2021/02/27 09:02:15  (): org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connection unexpectedly closed by remote task manager 'beam-loadtests-go-cogbk-flink-batch-82-w-2.c.apache-beam-testing.internal/10.128.0.186:42695'. This might indicate that the remote task manager was lost.
2021/02/27 09:02:16 Job state: FAILED
2021/02/27 09:02:16 Failed to execute job: job load0tests0go0flink0batch0cogbk0400227065603-root-0227085218-62380d27_e82a681f-b30f-4069-9f73-a7dd7e0158a3 failed
panic: Failed to execute job: job load0tests0go0flink0batch0cogbk0400227065603-root-0227085218-62380d27_e82a681f-b30f-4069-9f73-a7dd7e0158a3 failed

goroutine 1 [running]:
github.com/apache/beam/sdks/go/test/load/vendor/github.com/apache/beam/sdks/go/pkg/beam/log.Fatalf(0x11a2e00, 0xc00003e0b0, 0x1060143, 0x19, 0xc000393ee8, 0x1, 0x1)
	<https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/sdks/go/test/load/.gogradle/project_gopath/src/github.com/apache/beam/sdks/go/test/load/vendor/github.com/apache/beam/sdks/go/pkg/beam/log/log.go>:153 +0xec
main.main()
	<https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/sdks/go/test/load/.gogradle/project_gopath/src/github.com/apache/beam/sdks/go/test/load/cogbk/cogbk.go>:99 +0x577

> Task :sdks:go:test:load:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/sdks/go/test/load/build.gradle'> line: 65

* What went wrong:
Execution failed for task ':sdks:go:test:load:run'.
> Process 'command 'sh'' finished with non-zero exit value 2

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 10m 6s
6 actionable tasks: 6 executed

Publishing build scan...
https://gradle.com/s/m27mxvjbwgkds

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Go_CoGBK_Flink_batch #81

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/81/display/redirect?page=changes>

Changes:

[noreply] Minor fix in unit test

[samuelw] Add profiling option to nexmark gradle launching. Passing the

[Robert Bradshaw] [BEAM-11719] Use deterministic coders for grouping keys.

[Robert Bradshaw] Fix bigquery file load.

[Robert Bradshaw] Fix GroupWithCoderTest.test_basics_without_type_check to check

[Robert Bradshaw] Add the ability to register additional fallback coders.

[cvm] Minor bugfix in typecoders.

[Robert Bradshaw] [BEAM-11719] Pipeline option for non-deterministic key coders.

[Robert Bradshaw] Lint, import fix.

[samuelw] [BEAM-11727] Optimize ExecutionStateSampler by removing unnecessary

[dpcollins] [BEAM-10114] Fix PerSubscriptionPartitionSdf to not rely on the presence

[dpcollins] [BEAM-10114] Fix PerSubscriptionPartitionSdf to not rely on the presence

[Robert Bradshaw] Fixup

[samuelw] [BEAM-11730] Reduce context switching overhead for appliance reads by

[Fokko Driesprong] [BEAM-11864] Use objects.equals instead of raw comparison

[noreply] Refactor ZetaSqlDialectSpecTest and add some passing tests. (#14080)

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/thrift

[noreply] [BEAM-10961] enable strict dependency checking for sdks/java/io/parquet

[shehzaad] [BEAM-10961] enable strict dependency checking for

[noreply] [BEAM-10961] Enable strict dependency checking on

[Kenneth Knowles] Merge Fn API and runner v2 configurations for DataflowRunner

[noreply] [BEAM-9378] Add ignored tests which fail in various ways when querying

[noreply] [BEAM-11778] Create a wrapper for ZetaSQL catalog and refactor

[noreply] [BEAM-11870] Re-raise underlying exception for InvocationTargetException

[noreply] Puts more expensive BQ empty table check to the right of the 'and'


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-4 (beam) in workspace <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/>
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 248104ee800e3d7fd554ac48ba93c0f26f9b3952 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 248104ee800e3d7fd554ac48ba93c0f26f9b3952 # timeout=10
Commit message: "Merge pull request #13878:[BEAM-11730] Reduce context switching overhead for appliance reads by"
 > git rev-list --no-walk d967389e1f68f85f1b0e7665edc3a0b6ed3d70f1 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content 
JOB_SERVER_IMAGE=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
CLUSTER_NAME=beam-loadtests-go-cogbk-flink-batch-81
DETACHED_MODE=true
HARNESS_IMAGES_TO_PULL=gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest
FLINK_NUM_WORKERS=5
FLINK_DOWNLOAD_URL=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz
GCS_BUCKET=gs://beam-flink-cluster
HADOOP_DOWNLOAD_URL=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
FLINK_TASKMANAGER_SLOTS=1
ARTIFACTS_DIR=gs://beam-flink-cluster/beam-loadtests-go-cogbk-flink-batch-81
GCLOUD_ZONE=us-central1-a

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Go_CoGBK_Flink_batch] $ /bin/bash -xe /tmp/jenkins1141465857144382868.sh
+ echo Setting up flink cluster
Setting up flink cluster
[beam_LoadTests_Go_CoGBK_Flink_batch] $ /bin/bash -xe /tmp/jenkins7656310561181634855.sh
+ cd <https://ci-beam.apache.org/job/beam_LoadTests_Go_CoGBK_Flink_batch/ws/src/.test-infra/dataproc>
+ ./flink_cluster.sh create
+ GCLOUD_ZONE=us-central1-a
+ DATAPROC_VERSION=1.2
+ MASTER_NAME=beam-loadtests-go-cogbk-flink-batch-81-m
+ INIT_ACTIONS_FOLDER_NAME=init-actions
+ FLINK_INIT=gs://beam-flink-cluster/init-actions/flink.sh
+ BEAM_INIT=gs://beam-flink-cluster/init-actions/beam.sh
+ DOCKER_INIT=gs://beam-flink-cluster/init-actions/docker.sh
+ FLINK_LOCAL_PORT=8081
+ FLINK_TASKMANAGER_SLOTS=1
+ YARN_APPLICATION_MASTER=
+ create
+ upload_init_actions
+ echo 'Uploading initialization actions to GCS bucket: gs://beam-flink-cluster'
Uploading initialization actions to GCS bucket: gs://beam-flink-cluster
+ gsutil cp -r init-actions/beam.sh init-actions/docker.sh init-actions/flink.sh gs://beam-flink-cluster/init-actions
Copying file://init-actions/beam.sh [Content-Type=text/x-sh]...
/ [0 files][    0.0 B/  2.3 KiB]                                                / [1 files][  2.3 KiB/  2.3 KiB]                                                Copying file://init-actions/docker.sh [Content-Type=text/x-sh]...
/ [1 files][  2.3 KiB/  6.0 KiB]                                                / [2 files][  6.0 KiB/  6.0 KiB]                                                Copying file://init-actions/flink.sh [Content-Type=text/x-sh]...
/ [2 files][  6.0 KiB/ 13.5 KiB]                                                / [3 files][ 13.5 KiB/ 13.5 KiB]                                                
Operation completed over 3 objects/13.5 KiB.                                     
+ create_cluster
+ local metadata=flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,
+ metadata+=flink-start-yarn-session=true,
+ metadata+=flink-taskmanager-slots=1,
+ metadata+=hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest ]]
+ metadata+=,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest ]]
+ metadata+=,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
+ local image_version=1.2
+ echo 'Starting dataproc cluster. Dataproc version: 1.2'
Starting dataproc cluster. Dataproc version: 1.2
+ local num_dataproc_****s=6
+ gcloud dataproc clusters create beam-loadtests-go-cogbk-flink-batch-81 --region=global --num-****s=6 --initialization-actions gs://beam-flink-cluster/init-actions/docker.sh,gs://beam-flink-cluster/init-actions/beam.sh,gs://beam-flink-cluster/init-actions/flink.sh --metadata flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,flink-start-yarn-session=true,flink-taskmanager-slots=1,hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest, --image-version=1.2 --zone=us-central1-a --quiet
Waiting on operation [projects/apache-beam-testing/regions/global/operations/39c24c73-1993-34ef-bd26-c9227d4ba57d].
Waiting for cluster creation operation...
WARNING: For PD-Standard without local SSDs, we strongly recommend provisioning 1TB or larger to ensure consistently high I/O performance. See https://cloud.google.com/compute/docs/disks/performance for information on disk I/O performance.
..............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................done.
ERROR: (gcloud.dataproc.clusters.create) Operation [projects/apache-beam-testing/regions/global/operations/39c24c73-1993-34ef-bd26-c9227d4ba57d] failed: Multiple Errors:
 - Initialization action timed out. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/b6a92f00-f1a3-4c63-aa37-5bf511889fee/beam-loadtests-go-cogbk-flink-batch-81-m/dataproc-initialization-script-2_output
 - Initialization action timed out. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/b6a92f00-f1a3-4c63-aa37-5bf511889fee/beam-loadtests-go-cogbk-flink-batch-81-w-0/dataproc-initialization-script-2_output
 - Initialization action timed out. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/b6a92f00-f1a3-4c63-aa37-5bf511889fee/beam-loadtests-go-cogbk-flink-batch-81-w-1/dataproc-initialization-script-2_output
 - Initialization action timed out. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/b6a92f00-f1a3-4c63-aa37-5bf511889fee/beam-loadtests-go-cogbk-flink-batch-81-w-5/dataproc-initialization-script-2_output.
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org