You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/01/27 11:12:21 UTC

Build failed in Jenkins: beam_LoadTests_Go_GBK_Flink_Batch #64

See <https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/64/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-10925] Load SQL UDFs from jar.

[Kyle Weaver] Move JavaUdfLoader from zetasql -> sql.

[Kyle Weaver] Make JavaUdfDefinitions a subclass of JavaUdfLoader.

[Kyle Weaver] [BEAM-9541] Push all docker images from RC instead of hard-coding them.

[randomstep] [BEAM-9369] bump mockito-core to 3.7.7

[Fokko Driesprong] BEAM-4986: Bump to Apache Parquet 1.11.1

[Kyle Weaver] Don't set context classloader.

[Pablo Estrada] Revert "Revert "Merge pull request #12647 from [BEAM-10378]

[Pablo Estrada] Fixing checker framework checks

[Kyle Weaver] [BEAM-9541] Update Python SDK's Flink version list.

[Kyle Weaver] [BEAM-9541] Create a Gradle task to push all docker images, and use it

[Kyle Weaver] [BEAM-11689] Use public.nexus.pentaho.org for pentaho dependencies

[noreply] [BEAM-11272] Remove combiner label constructor arg (#13355)

[noreply] [BEAM-11691] Skip JavaUdfLoaderTest instead of failing when jar path

[noreply] Merge pull request #13757: [BEAM-11640] Linkage Checker version upgrade


------------------------------------------
[...truncated 116.55 KB...]
  >
  coders: <
    key: "c3"
    value: <
      spec: <
        urn: "beam:coder:iterable:v1"
      >
      component_coder_ids: "c0"
    >
  >
  coders: <
    key: "c4"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c3"
    >
  >
  coders: <
    key: "c5"
    value: <
      spec: <
        urn: "beam:go:coder:custom:v1"
        payload: "CgRqc29uEosCCBUaKAoLTnVtRWxlbWVudHMaAggCIhJqc29uOiJudW1fcmVjb3JkcyIyAQAaLwoNSW5pdGlhbFNwbGl0cxoCCAIiFWpzb246ImluaXRpYWxfc3BsaXRzIigIMgEBGiMKB0tleVNpemUaAggCIg9qc29uOiJrZXlfc2l6ZSIoEDIBAhonCglWYWx1ZVNpemUaAggCIhFqc29uOiJ2YWx1ZV9zaXplIigYMgEDGioKCk51bUhvdEtleXMaAggCIhNqc29uOiJudW1faG90X2tleXMiKCAyAQQaMgoOSG90S2V5RnJhY3Rpb24aAggOIhdqc29uOiJob3Rfa2V5X2ZyYWN0aW9uIigoMgEFGnkKX2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L2xvYWQvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS5qc29uRW5jEhYIFiIECBlADyoGCBQSAggIKgQIGUABIn8KX2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L2xvYWQvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS5qc29uRGVjEhwIFiIECBlAAyIGCBQSAggIKgQIGUAPKgQIGUAB"
      >
    >
  >
  coders: <
    key: "c6"
    value: <
      spec: <
        urn: "beam:coder:length_prefix:v1"
      >
      component_coder_ids: "c5"
    >
  >
  coders: <
    key: "c7"
    value: <
      spec: <
        urn: "beam:go:coder:custom:v1"
        payload: "ChdvZmZzZXRyYW5nZS5SZXN0cmljdGlvbhKAAQgaSnxnaXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vdGVzdC9sb2FkL3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vaW8vcnRyYWNrZXJzL29mZnNldHJhbmdlLlJlc3RyaWN0aW9uGpACCnhnaXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vdGVzdC9sb2FkL3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vaW8vcnRyYWNrZXJzL29mZnNldHJhbmdlLnJlc3RFbmMSkwEIFiKAAQgaSnxnaXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vdGVzdC9sb2FkL3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vaW8vcnRyYWNrZXJzL29mZnNldHJhbmdlLlJlc3RyaWN0aW9uKgYIFBICCAgqBAgZQAEikAIKeGdpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L2xvYWQvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS9pby9ydHJhY2tlcnMvb2Zmc2V0cmFuZ2UucmVzdERlYxKTAQgWIgYIFBICCAgqgAEIGkp8Z2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvbG9hZC92ZW5kb3IvZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3BrZy9iZWFtL2lvL3J0cmFja2Vycy9vZmZzZXRyYW5nZS5SZXN0cmljdGlvbioECBlAAQ=="
      >
    >
  >
  coders: <
    key: "c8"
    value: <
      spec: <
        urn: "beam:coder:length_prefix:v1"
      >
      component_coder_ids: "c7"
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n>gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_****_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006****"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "e6"
root_transform_ids: "e7"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2021/01/27 11:11:40 Prepared job with id: load-tests-go-flink-batch-gbk-3-0127065404_a7aa3ca2-8667-4297-9aef-8f5d8ea4d1c6 and staging token: load-tests-go-flink-batch-gbk-3-0127065404_a7aa3ca2-8667-4297-9aef-8f5d8ea4d1c6
2021/01/27 11:11:40 Using specified **** binary: 'linux_amd64/group_by_key'
2021/01/27 11:11:42 Staged binary artifact with token: 
2021/01/27 11:11:42 Submitted job: load0tests0go0flink0batch0gbk0300127065404-root-0127111142-16232cac_ecc563dc-1ea0-4b6f-b14a-62c28b6da7b4
2021/01/27 11:11:42 Job state: STOPPED
2021/01/27 11:11:42 Job state: STARTING
2021/01/27 11:11:42 Job state: RUNNING
2021/01/27 11:12:20  (): java.util.concurrent.ExecutionException: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 10f6783b295c8cef0179e84f59270bf1)
	at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
	at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
	at org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:864)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:199)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:118)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:85)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 10f6783b295c8cef0179e84f59270bf1)
	at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$null$6(ClusterClientJobClientAdapter.java:112)
	at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
	at java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
	at org.apache.flink.client.program.rest.RestClusterClient.lambda$pollResourceAsync$21(RestClusterClient.java:565)
	at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
	at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
	at org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:291)
	at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
	at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
	at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
	at java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:575)
	at java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:943)
	at java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456)
	... 3 more
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
	at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$null$6(ClusterClientJobClientAdapter.java:110)
	... 19 more
Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy
	at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:110)
	at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:76)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:192)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:186)
	at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:180)
	at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:496)
	at org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:380)
	at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:284)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:199)
	at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
	at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
	at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
	at akka.actor.ActorCell.invoke(ActorCell.scala:561)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
	at akka.dispatch.Mailbox.run(Mailbox.scala:225)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
	at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.Exception: The data preparation for task 'GroupReduce (GroupReduce at CoGBK)' , caused an error: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Lost connection to task manager 'beam-loadtests-go-gbk-flink-batch-64-w-3.c.apache-beam-testing.internal/10.128.0.46:33435'. This indicates that the remote task manager was lost.
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480)
	at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
	at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:708)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:533)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.RuntimeException: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Lost connection to task manager 'beam-loadtests-go-gbk-flink-batch-64-w-3.c.apache-beam-testing.internal/10.128.0.46:33435'. This indicates that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger.getIterator(UnilateralSortMerger.java:650)
	at org.apache.flink.runtime.operators.BatchTask.getInput(BatchTask.java:1110)
	at org.apache.flink.runtime.operators.GroupReduceDriver.prepare(GroupReduceDriver.java:99)
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:474)
	... 4 more
Caused by: java.io.IOException: Thread 'SortMerger Reading Thread' terminated due to an exception: Lost connection to task manager 'beam-loadtests-go-gbk-flink-batch-64-w-3.c.apache-beam-testing.internal/10.128.0.46:33435'. This indicates that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger$ThreadBase.run(UnilateralSortMerger.java:831)
Caused by: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Lost connection to task manager 'beam-loadtests-go-gbk-flink-batch-64-w-3.c.apache-beam-testing.internal/10.128.0.46:33435'. This indicates that the remote task manager was lost.
	at org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.exceptionCaught(CreditBasedPartitionRequestClientHandler.java:160)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:297)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:276)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireExceptionCaught(AbstractChannelHandlerContext.java:268)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.exceptionCaught(DefaultChannelPipeline.java:1388)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:297)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:276)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireExceptionCaught(DefaultChannelPipeline.java:918)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.handleReadException(AbstractNioByteChannel.java:125)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:174)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:697)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:632)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:549)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:511)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:918)
	at org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: Connection reset by peer
	at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
	at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)
	at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
	at sun.nio.ch.IOUtil.read(IOUtil.java:192)
	at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:377)
	at org.apache.flink.shaded.netty4.io.netty.buffer.PooledByteBuf.setBytes(PooledByteBuf.java:247)
	at org.apache.flink.shaded.netty4.io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:1140)
	at org.apache.flink.shaded.netty4.io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:347)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:148)
	... 7 more
2021/01/27 11:12:20  (): java.io.IOException: Connection reset by peer
2021/01/27 11:12:20 Job state: FAILED
2021/01/27 11:12:20 Failed to execute job: job load0tests0go0flink0batch0gbk0300127065404-root-0127111142-16232cac_ecc563dc-1ea0-4b6f-b14a-62c28b6da7b4 failed
panic: Failed to execute job: job load0tests0go0flink0batch0gbk0300127065404-root-0127111142-16232cac_ecc563dc-1ea0-4b6f-b14a-62c28b6da7b4 failed

goroutine 1 [running]:
github.com/apache/beam/sdks/go/test/load/vendor/github.com/apache/beam/sdks/go/pkg/beam/log.Fatalf(0x11997e0, 0xc00003e0c0, 0x1057ca0, 0x19, 0xc0000ddee8, 0x1, 0x1)
	<https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/ws/src/sdks/go/test/load/.gogradle/project_gopath/src/github.com/apache/beam/sdks/go/test/load/vendor/github.com/apache/beam/sdks/go/pkg/beam/log/log.go>:153 +0xec
main.main()
	<https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/ws/src/sdks/go/test/load/.gogradle/project_gopath/src/github.com/apache/beam/sdks/go/test/load/group_by_key/group_by_key.go>:82 +0x47e

> Task :sdks:go:test:load:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/ws/src/sdks/go/test/load/build.gradle'> line: 65

* What went wrong:
Execution failed for task ':sdks:go:test:load:run'.
> Process 'command 'sh'' finished with non-zero exit value 2

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 46s
6 actionable tasks: 6 executed

Publishing build scan...
https://gradle.com/s/2ylp77w2xtees

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_LoadTests_Go_GBK_Flink_Batch #66

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/66/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Go_GBK_Flink_Batch #65

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/65/display/redirect?page=changes>

Changes:

[sonam.ramchand] Enabled strict dependency on flink runner

[sonam.ramchand] Added new line

[sonam.ramchand] removed checker-qual

[sonam.ramchand] Removed flink-clients dep

[sonam.ramchand] set enableStrictDependencies:true

[sonam.ramchand] made flink-clients runtimeOnly

[noreply] Remove stray colon.

[zyichi] [BEAM-11679] Override PubsubUnboundedSource transform for dataflow

[zyichi] Populate PubsubMessage message id

[zyichi] Enable tests

[zyichi] Exclude failing FhirIO ITs

[zyichi] Include runner v2 IT tests in java post commit

[Ismaël Mejía] [BEAM-11697] Upgrade Flink runner to Flink versions 1.12.1 and 1.11.3

[shehzaad] strict dependency checking on sdks/io/amazon-web-services*

[shehzaad] move amazon-web-services* changes to another PR

[shehzaad] undo previous erroneous commit

[shehzaad] better use of variables

[randomstep] [BEAM-8725] bump snappy-java to 1.1.8.4

[noreply] [BEAM-10961] Enable strict dependency checking on

[nielm] Add BigDecimal support for SpannerIO

[noreply] [BEAM-11695] Remove translations.pack_combiners from default optimizers

[zyichi] [BEAM-11699] Correct python environment directly for fnApiJobServerSetup

[Kyle Weaver] [BEAM-11689] Add public.nexus.pentaho.org to offline repositories.

[Chamikara Madhusanka Jayalath] Moving to 2.29.0-SNAPSHOT on master branch.

[noreply] [BEAM-11531] Add pd.to_datetime, handle DeferredBase args in

[Andrew Pilloud] [BEAM-11165] Async ZetaSQL Calc

[noreply] [BEAM-11695] Combiner packing in Dataflow (#13763)

[noreply] Merge pull request #13496 from [BEAM-11408] Integrate BigQuery sink


------------------------------------------
[...truncated 136.39 KB...]
      windowing_strategy_id: "w0"
    >
  >
  pcollections: <
    key: "n5"
    value: <
      unique_name: "n5"
      coder_id: "c4"
      is_bounded: BOUNDED
      windowing_strategy_id: "w0"
    >
  >
  pcollections: <
    key: "n6"
    value: <
      unique_name: "n6"
      coder_id: "c1"
      is_bounded: BOUNDED
      windowing_strategy_id: "w0"
    >
  >
  pcollections: <
    key: "n7"
    value: <
      unique_name: "n7"
      coder_id: "c1"
      is_bounded: BOUNDED
      windowing_strategy_id: "w0"
    >
  >
  windowing_strategies: <
    key: "w0"
    value: <
      window_fn: <
        urn: "beam:window_fn:global_windows:v1"
      >
      merge_status: NON_MERGING
      window_coder_id: "c2"
      trigger: <
        default: <
        >
      >
      accumulation_mode: DISCARDING
      output_time: END_OF_WINDOW
      closing_behavior: EMIT_IF_NONEMPTY
      OnTimeBehavior: FIRE_ALWAYS
      environment_id: "go"
    >
  >
  coders: <
    key: "c0"
    value: <
      spec: <
        urn: "beam:coder:bytes:v1"
      >
    >
  >
  coders: <
    key: "c1"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c0"
    >
  >
  coders: <
    key: "c2"
    value: <
      spec: <
        urn: "beam:coder:global_window:v1"
      >
    >
  >
  coders: <
    key: "c3"
    value: <
      spec: <
        urn: "beam:coder:iterable:v1"
      >
      component_coder_ids: "c0"
    >
  >
  coders: <
    key: "c4"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "c0"
      component_coder_ids: "c3"
    >
  >
  coders: <
    key: "c5"
    value: <
      spec: <
        urn: "beam:go:coder:custom:v1"
        payload: "CgRqc29uEosCCBUaKAoLTnVtRWxlbWVudHMaAggCIhJqc29uOiJudW1fcmVjb3JkcyIyAQAaLwoNSW5pdGlhbFNwbGl0cxoCCAIiFWpzb246ImluaXRpYWxfc3BsaXRzIigIMgEBGiMKB0tleVNpemUaAggCIg9qc29uOiJrZXlfc2l6ZSIoEDIBAhonCglWYWx1ZVNpemUaAggCIhFqc29uOiJ2YWx1ZV9zaXplIigYMgEDGioKCk51bUhvdEtleXMaAggCIhNqc29uOiJudW1faG90X2tleXMiKCAyAQQaMgoOSG90S2V5RnJhY3Rpb24aAggOIhdqc29uOiJob3Rfa2V5X2ZyYWN0aW9uIigoMgEFGnkKX2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L2xvYWQvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS5qc29uRW5jEhYIFiIECBlADyoGCBQSAggIKgQIGUABIn8KX2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L2xvYWQvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS5qc29uRGVjEhwIFiIECBlAAyIGCBQSAggIKgQIGUAPKgQIGUAB"
      >
    >
  >
  coders: <
    key: "c6"
    value: <
      spec: <
        urn: "beam:coder:length_prefix:v1"
      >
      component_coder_ids: "c5"
    >
  >
  coders: <
    key: "c7"
    value: <
      spec: <
        urn: "beam:go:coder:custom:v1"
        payload: "ChdvZmZzZXRyYW5nZS5SZXN0cmljdGlvbhKAAQgaSnxnaXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vdGVzdC9sb2FkL3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vaW8vcnRyYWNrZXJzL29mZnNldHJhbmdlLlJlc3RyaWN0aW9uGpACCnhnaXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vdGVzdC9sb2FkL3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vaW8vcnRyYWNrZXJzL29mZnNldHJhbmdlLnJlc3RFbmMSkwEIFiKAAQgaSnxnaXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vdGVzdC9sb2FkL3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vaW8vcnRyYWNrZXJzL29mZnNldHJhbmdlLlJlc3RyaWN0aW9uKgYIFBICCAgqBAgZQAEikAIKeGdpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L2xvYWQvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS9pby9ydHJhY2tlcnMvb2Zmc2V0cmFuZ2UucmVzdERlYxKTAQgWIgYIFBICCAgqgAEIGkp8Z2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvbG9hZC92ZW5kb3IvZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3BrZy9iZWFtL2lvL3J0cmFja2Vycy9vZmZzZXRyYW5nZS5SZXN0cmljdGlvbioECBlAAQ=="
      >
    >
  >
  coders: <
    key: "c8"
    value: <
      spec: <
        urn: "beam:coder:length_prefix:v1"
      >
      component_coder_ids: "c7"
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n>gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_****_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006****"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "e6"
root_transform_ids: "e7"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2021/01/28 11:14:10 Prepared job with id: load-tests-go-flink-batch-gbk-6-0128065419_b2c5808d-18ac-47bc-b1d4-657c4b63dc63 and staging token: load-tests-go-flink-batch-gbk-6-0128065419_b2c5808d-18ac-47bc-b1d4-657c4b63dc63
2021/01/28 11:14:10 Using specified **** binary: 'linux_amd64/group_by_key'
2021/01/28 11:14:12 Staged binary artifact with token: 
2021/01/28 11:14:12 Submitted job: load0tests0go0flink0batch0gbk0600128065419-root-0128111412-41cf7951_b34182ff-6862-4685-a2ac-e85efe9f7c85
2021/01/28 11:14:12 Job state: STOPPED
2021/01/28 11:14:12 Job state: STARTING
2021/01/28 11:14:12 Job state: RUNNING
2021/01/28 11:24:26 Job state: DONE
2021/01/28 11:24:26 Failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"n2/SplitAndSize0"}
2021/01/28 11:24:26 Failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\xa3\x01" labels:{key:"PCOLLECTION" value:"n5"}
2021/01/28 11:24:26 Failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"n1"}
2021/01/28 11:24:26 Failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"n2/SplitAndSize0"}
2021/01/28 11:24:26 Failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\xa3\x01" labels:{key:"PCOLLECTION" value:"n5"}
2021/01/28 11:24:26 Failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"n1"}

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 10m 21s
6 actionable tasks: 5 executed, 1 up-to-date

Publishing build scan...
https://gradle.com/s/6r3ipg3ciak66

[beam_LoadTests_Go_GBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins2452092990298096631.sh
+ echo Changing number of ****s to 16
Changing number of ****s to 16
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content 
FLINK_NUM_WORKERS=16

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Go_GBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins1043317783272675307.sh
+ cd <https://ci-beam.apache.org/job/beam_LoadTests_Go_GBK_Flink_Batch/ws/src/.test-infra/dataproc>
+ ./flink_cluster.sh restart
+ GCLOUD_ZONE=us-central1-a
+ DATAPROC_VERSION=1.2
+ MASTER_NAME=beam-loadtests-go-gbk-flink-batch-65-m
+ INIT_ACTIONS_FOLDER_NAME=init-actions
+ FLINK_INIT=gs://beam-flink-cluster/init-actions/flink.sh
+ BEAM_INIT=gs://beam-flink-cluster/init-actions/beam.sh
+ DOCKER_INIT=gs://beam-flink-cluster/init-actions/docker.sh
+ FLINK_LOCAL_PORT=8081
+ FLINK_TASKMANAGER_SLOTS=1
+ YARN_APPLICATION_MASTER=
+ restart
+ delete
+ gcloud dataproc clusters delete beam-loadtests-go-gbk-flink-batch-65 --region=global --quiet
Waiting on operation [projects/apache-beam-testing/regions/global/operations/359e35e2-7fd1-3a38-b6c1-0a4249ff7633].
Waiting for cluster deletion operation...
...................................done.
Deleted [https://dataproc.googleapis.com/v1/projects/apache-beam-testing/regions/global/clusters/beam-loadtests-go-gbk-flink-batch-65].
+ create
+ upload_init_actions
+ echo 'Uploading initialization actions to GCS bucket: gs://beam-flink-cluster'
Uploading initialization actions to GCS bucket: gs://beam-flink-cluster
+ gsutil cp -r init-actions/beam.sh init-actions/docker.sh init-actions/flink.sh gs://beam-flink-cluster/init-actions
Copying file://init-actions/beam.sh [Content-Type=text/x-sh]...
/ [0 files][    0.0 B/  2.3 KiB]                                                / [1 files][  2.3 KiB/  2.3 KiB]                                                Copying file://init-actions/docker.sh [Content-Type=text/x-sh]...
/ [1 files][  2.3 KiB/  6.0 KiB]                                                / [2 files][  6.0 KiB/  6.0 KiB]                                                Copying file://init-actions/flink.sh [Content-Type=text/x-sh]...
/ [2 files][  6.0 KiB/ 13.5 KiB]                                                / [3 files][ 13.5 KiB/ 13.5 KiB]                                                
Operation completed over 3 objects/13.5 KiB.                                     
+ create_cluster
+ local metadata=flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,
+ metadata+=flink-start-yarn-session=true,
+ metadata+=flink-taskmanager-slots=1,
+ metadata+=hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest ]]
+ metadata+=,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest ]]
+ metadata+=,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
+ local image_version=1.2
+ echo 'Starting dataproc cluster. Dataproc version: 1.2'
Starting dataproc cluster. Dataproc version: 1.2
+ local num_dataproc_****s=17
+ gcloud dataproc clusters create beam-loadtests-go-gbk-flink-batch-65 --region=global --num-****s=17 --initialization-actions gs://beam-flink-cluster/init-actions/docker.sh,gs://beam-flink-cluster/init-actions/beam.sh,gs://beam-flink-cluster/init-actions/flink.sh --metadata flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,flink-start-yarn-session=true,flink-taskmanager-slots=1,hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_go_sdk:latest,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest, --image-version=1.2 --zone=us-central1-a --quiet
Waiting on operation [projects/apache-beam-testing/regions/global/operations/9613989e-681f-3d68-9d93-8da6ca593b6b].
Waiting for cluster creation operation...
WARNING: For PD-Standard without local SSDs, we strongly recommend provisioning 1TB or larger to ensure consistently high I/O performance. See https://cloud.google.com/compute/docs/disks/performance for information on disk I/O performance.
................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................done.
ERROR: (gcloud.dataproc.clusters.create) Operation [projects/apache-beam-testing/regions/global/operations/9613989e-681f-3d68-9d93-8da6ca593b6b] failed: Multiple Errors:
 - Initialization action timed out. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/681b84d5-9670-4056-b84d-4a5741165700/beam-loadtests-go-gbk-flink-batch-65-w-1/dataproc-initialization-script-2_output
 - Initialization action timed out. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/681b84d5-9670-4056-b84d-4a5741165700/beam-loadtests-go-gbk-flink-batch-65-w-16/dataproc-initialization-script-2_output.
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org