You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/12/07 18:37:45 UTC

Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #135

See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/135/display/redirect?page=changes>

Changes:

[robertwb] [BEAM-8882] Implement Impulse() for BundleBasedRunner.

[robertwb] [BEAM-8882] Make Create fn-api agnostic.

[robertwb] [BEAM-8882] Fully specify types for Create composite.

[robertwb] [BEAM-8882] Make Read fn-api agnostic.

[robertwb] [BEAM-8882] Cleanup always-on use_sdf_bounded_source option.

[robertwb] [BEAM-8882] Annotate ParDo and CombineValues operations with proto

[robertwb] [BEAM-8882] Unconditionally populate pipeline_proto_coder_id.

[kcweaver] [BEAM-8835] Stage artifacts to BEAM-PIPELINE dir in zip

[kcweaver] [BEAM-8835] Check for leading slash in zip file paths.

[robertwb] [BEAM-8882] Fix overly-sensitive tests.

[robertwb] Fix sdf tests from create.

[robertwb] [BEAM-8882] Avoid attaching unrecognized properties.

[robertwb] [BEAM-8882] Accommodations for JRH.

[robertwb] Minor cleanup.

[github] [BEAM-8882] Fully populate log messages. (#10292)

[github] [BEAM-3865] Stronger trigger tests. (#10192)

[pabloem] Merge pull request #10236 from [BEAM-8335] Add method to

[bhulette] [BEAM-8427] Add MongoDB to SQL documentation (#10273)


------------------------------------------
[...truncated 215.33 KB...]
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:flink_version:v1"
    value {
      string_value: "1.9"
    }
  }
  fields {
    key: "beam:option:gcs_performance_metrics:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:job_endpoint:v1"
    value {
      string_value: "localhost:8099"
    }
  }
  fields {
    key: "beam:option:job_name:v1"
    value {
      string_value: "load-tests-python-flink-batch-cogbk-3-1207150519"
    }
  }
  fields {
    key: "beam:option:job_port:v1"
    value {
      string_value: "0"
    }
  }
  fields {
    key: "beam:option:job_server_timeout:v1"
    value {
      string_value: "60"
    }
  }
  fields {
    key: "beam:option:no_auth:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:object_reuse:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:parallelism:v1"
    value {
      string_value: "5"
    }
  }
  fields {
    key: "beam:option:pipeline_type_check:v1"
    value {
      bool_value: true
    }
  }
  fields {
    key: "beam:option:profile_cpu:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_memory:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_sample_rate:v1"
    value {
      number_value: 1.0
    }
  }
  fields {
    key: "beam:option:project:v1"
    value {
      string_value: "apache-beam-testing"
    }
  }
  fields {
    key: "beam:option:retain_docker_containers:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:retain_externalized_checkpoints_on_cancellation:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:runtime_type_check:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:save_main_session:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:sdk_location:v1"
    value {
      string_value: "container"
    }
  }
  fields {
    key: "beam:option:sdk_worker_parallelism:v1"
    value {
      string_value: "1"
    }
  }
  fields {
    key: "beam:option:shutdown_sources_on_final_watermark:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:streaming:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:temp_location:v1"
    value {
      string_value: "gs://temp-storage-for-perf-tests/loadtests"
    }
  }
  fields {
    key: "beam:option:type_check_strictness:v1"
    value {
      string_value: "DEFAULT_TO_ANY"
    }
  }
  fields {
    key: "beam:option:update:v1"
    value {
      bool_value: false
    }
  }
}
job_name: "job"

apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: org.apache.flink.client.program.ProgramInvocationException: Job failed. (JobID: c51a14d849559430ee113b93d90c6ce4)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:301)
	at org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:209)
	at org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:186)
	at org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:173)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:191)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:116)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:84)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:81)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
	... 16 more
Caused by: java.util.concurrent.TimeoutException: Heartbeat of TaskManager with id container_e01_1575735438298_0001_01_000004 timed out.
	at org.apache.flink.runtime.jobmaster.JobMaster$TaskManagerHeartbeatListener.notifyHeartbeatTimeout(JobMaster.java:1149)
	at org.apache.flink.runtime.heartbeat.HeartbeatMonitorImpl.run(HeartbeatMonitorImpl.java:109)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:397)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:190)
	at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
	at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
	at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
	at akka.actor.ActorCell.invoke(ActorCell.scala:561)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
	at akka.dispatch.Mailbox.run(Mailbox.scala:225)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
	at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

root: ERROR: java.util.concurrent.TimeoutException: Heartbeat of TaskManager with id container_e01_1575735438298_0001_01_000004 timed out.
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 1646.583s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 55

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 27m 31s
3 actionable tasks: 2 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/346lti54gvr42

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_LoadTests_Python_coGBK_Flink_Batch #144

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/144/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #143

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/143/display/redirect>

Changes:


------------------------------------------
[...truncated 247.30 KB...]
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:flink_version:v1"
    value {
      string_value: "1.9"
    }
  }
  fields {
    key: "beam:option:gcs_performance_metrics:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:job_endpoint:v1"
    value {
      string_value: "localhost:8099"
    }
  }
  fields {
    key: "beam:option:job_name:v1"
    value {
      string_value: "load-tests-python-flink-batch-cogbk-4-1215150459"
    }
  }
  fields {
    key: "beam:option:job_port:v1"
    value {
      string_value: "0"
    }
  }
  fields {
    key: "beam:option:job_server_timeout:v1"
    value {
      string_value: "60"
    }
  }
  fields {
    key: "beam:option:no_auth:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:object_reuse:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:parallelism:v1"
    value {
      string_value: "5"
    }
  }
  fields {
    key: "beam:option:pipeline_type_check:v1"
    value {
      bool_value: true
    }
  }
  fields {
    key: "beam:option:profile_cpu:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_memory:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_sample_rate:v1"
    value {
      number_value: 1.0
    }
  }
  fields {
    key: "beam:option:project:v1"
    value {
      string_value: "apache-beam-testing"
    }
  }
  fields {
    key: "beam:option:retain_docker_containers:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:retain_externalized_checkpoints_on_cancellation:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:runtime_type_check:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:save_main_session:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:sdk_location:v1"
    value {
      string_value: "container"
    }
  }
  fields {
    key: "beam:option:sdk_worker_parallelism:v1"
    value {
      string_value: "1"
    }
  }
  fields {
    key: "beam:option:shutdown_sources_on_final_watermark:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:streaming:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:temp_location:v1"
    value {
      string_value: "gs://temp-storage-for-perf-tests/loadtests"
    }
  }
  fields {
    key: "beam:option:type_check_strictness:v1"
    value {
      string_value: "DEFAULT_TO_ANY"
    }
  }
  fields {
    key: "beam:option:update:v1"
    value {
      bool_value: false
    }
  }
}
job_name: "job"

apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: org.apache.flink.client.program.ProgramInvocationException: Job failed. (JobID: 48a4c6dcb97f806c842078eb31c738b5)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:301)
	at org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:209)
	at org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:186)
	at org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:173)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:191)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:116)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:84)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:81)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
	... 16 more
Caused by: java.util.concurrent.TimeoutException: Heartbeat of TaskManager with id container_e01_1576426569989_0001_01_000006 timed out.
	at org.apache.flink.runtime.jobmaster.JobMaster$TaskManagerHeartbeatListener.notifyHeartbeatTimeout(JobMaster.java:1149)
	at org.apache.flink.runtime.heartbeat.HeartbeatMonitorImpl.run(HeartbeatMonitorImpl.java:109)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:397)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:190)
	at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
	at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
	at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
	at akka.actor.ActorCell.invoke(ActorCell.scala:561)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
	at akka.dispatch.Mailbox.run(Mailbox.scala:225)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
	at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

root: ERROR: java.util.concurrent.TimeoutException: Heartbeat of TaskManager with id container_e01_1576426569989_0001_01_000006 timed out.
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 1894.514s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 55

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 31m 39s
3 actionable tasks: 2 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/eeshnbx2sbp5g

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #142

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/142/display/redirect?page=changes>

Changes:

[ehudm] [BEAM-3713] pytest migration: py3x-{gcp,cython}

[younghoono] [BEAM-7970] Improved error help in Go PROTOBUF.md

[robertwb] Fixed instructions and the example of running Interactive Beam on Flink.


------------------------------------------
[...truncated 189.10 KB...]
      string_value: "1.9"
    }
  }
  fields {
    key: "beam:option:gcs_performance_metrics:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:job_endpoint:v1"
    value {
      string_value: "localhost:8099"
    }
  }
  fields {
    key: "beam:option:job_name:v1"
    value {
      string_value: "load-tests-python-flink-batch-cogbk-2-1214150456"
    }
  }
  fields {
    key: "beam:option:job_port:v1"
    value {
      string_value: "0"
    }
  }
  fields {
    key: "beam:option:job_server_timeout:v1"
    value {
      string_value: "60"
    }
  }
  fields {
    key: "beam:option:no_auth:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:object_reuse:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:parallelism:v1"
    value {
      string_value: "5"
    }
  }
  fields {
    key: "beam:option:pipeline_type_check:v1"
    value {
      bool_value: true
    }
  }
  fields {
    key: "beam:option:profile_cpu:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_memory:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_sample_rate:v1"
    value {
      number_value: 1.0
    }
  }
  fields {
    key: "beam:option:project:v1"
    value {
      string_value: "apache-beam-testing"
    }
  }
  fields {
    key: "beam:option:retain_docker_containers:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:retain_externalized_checkpoints_on_cancellation:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:runtime_type_check:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:save_main_session:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:sdk_location:v1"
    value {
      string_value: "container"
    }
  }
  fields {
    key: "beam:option:sdk_worker_parallelism:v1"
    value {
      string_value: "1"
    }
  }
  fields {
    key: "beam:option:shutdown_sources_on_final_watermark:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:streaming:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:temp_location:v1"
    value {
      string_value: "gs://temp-storage-for-perf-tests/loadtests"
    }
  }
  fields {
    key: "beam:option:type_check_strictness:v1"
    value {
      string_value: "DEFAULT_TO_ANY"
    }
  }
  fields {
    key: "beam:option:update:v1"
    value {
      bool_value: false
    }
  }
}
job_name: "job"

apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: org.apache.flink.client.program.ProgramInvocationException: Job failed. (JobID: 4a0904d409266a96757b6207c02bfa0e)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:301)
	at org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:209)
	at org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:186)
	at org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:173)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:191)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:116)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:84)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:81)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
	... 16 more
Caused by: java.lang.Exception: The data preparation for task 'GroupReduce (GroupReduce at CoGroupByKey /GroupByKey)' , caused an error: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-142-w-2.c.apache-beam-testing.internal/10.128.0.51:36443'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480)
	at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
	at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
	... 1 more
Caused by: java.lang.RuntimeException: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-142-w-2.c.apache-beam-testing.internal/10.128.0.51:36443'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger.getIterator(UnilateralSortMerger.java:650)
	at org.apache.flink.runtime.operators.BatchTask.getInput(BatchTask.java:1109)
	at org.apache.flink.runtime.operators.GroupReduceDriver.prepare(GroupReduceDriver.java:99)
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:474)
	... 4 more
Caused by: java.io.IOException: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-142-w-2.c.apache-beam-testing.internal/10.128.0.51:36443'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger$ThreadBase.run(UnilateralSortMerger.java:831)
Caused by: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-142-w-2.c.apache-beam-testing.internal/10.128.0.51:36443'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.channelInactive(CreditBasedPartitionRequestClientHandler.java:136)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
	at org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInputClosed(ByteToMessageDecoder.java:390)
	at org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInactive(ByteToMessageDecoder.java:355)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1429)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:947)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:826)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:404)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:474)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:909)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-142-w-2.c.apache-beam-testing.internal/10.128.0.51:36443'. This might indicate that the remote task manager was lost.
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 1719.896s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 55

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 28m 47s
3 actionable tasks: 2 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/brfadtl47yckq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #141

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/141/display/redirect?page=changes>

Changes:

[suztomo] [BEAM-8911] New Guava version: 25.1-jre

[zyichi] [BEAM-8886] Add a mongodb io dataflow integration test

[heejong] [BEAM-8905] matching Java PCollectionTuple translation naming convention

[michal.walenia] [BEAM-8947] Change Flink docker image names in Jenkins jobs to match

[github] [BEAM-8786] Fixes a link in readme

[pawel.pasterz] [BEAM-8946] Publish collection size of data written during MongoDBIOIT

[chamikara] Merge pull request #10347: [BEAM-8885] PubsubGrpcClient doesn't respect

[kenn] [BEAM-8917] jsr305 dependency declaration for Nullable class (#10324)

[xinyuliu.us] [BEAM-8342]: upgrade to samza 1.3.0 (#10357)

[aaltay] Make model_pcollection snippet self-contained (#10343)

[pabloem] Merge pull request #10050 from [BEAM-8575] Add streaming test case for

[github] Run beam_CancelStaleDataflowJobs every 4 hours.

[tvalentyn] [BEAM-8575] Added a unit test to test Combine works with sessions.

[github] [GoSDK] Improve StateChannel resilience. (#10363)

[tweise] [BEAM-8273] Expand portability environment documentation (#10116)


------------------------------------------
[...truncated 37.43 KB...]
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_artifact_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_expansion_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_expansion_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_fn_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_fn_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_interactive_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_job_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_job_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_provision_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_provision_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_runner_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_runner_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/endpoints_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/external_transforms_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/metrics_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/schema_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/standard_window_fns_pb2.py>
warning: no files found matching 'README.md'
warning: no files found matching 'NOTICE'
warning: no files found matching 'LICENSE'
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md


> Task :sdks:python:container:py2:copyDockerfileDependencies

> Task :sdks:go:resolveBuildDependencies
Resolving google.golang.org/api: commit='386d4e5f4f92f86e6aec85985761bba4b938a2d5', urls=[https://code.googlesource.com/google-api-go-client]
Resolving google.golang.org/genproto: commit='2b5a72b8730b0b16380010cfe5286c42108d88e7', urls=[https://github.com/google/go-genproto]
Resolving google.golang.org/grpc: commit='7646b5360d049a7ca31e9133315db43456f39e2e', urls=[https://github.com/grpc/grpc-go]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]

> Task :sdks:go:installDependencies
> Task :sdks:go:buildLinuxAmd64
> Task :sdks:go:goBuild

> Task :sdks:python:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/go>

> Task :sdks:python:container:installDependencies
> Task :sdks:python:container:buildDarwinAmd64
> Task :sdks:python:container:buildLinuxAmd64
> Task :sdks:python:container:goBuild
> Task :sdks:python:container:py2:copyLauncherDependencies
> Task :sdks:python:container:py2:dockerPrepare
> Task :sdks:python:container:py2:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 6m 16s
18 actionable tasks: 17 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/fez7fdjvupw7o

Build step 'Invoke Gradle script' changed build result to SUCCESS
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins490318573414196720.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins1737597982841545146.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/python2.7_sdk gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins3893312599434076003.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins5522599987451354138.sh
+ docker push gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
The push refers to repository [gcr.io/apache-beam-testing/beam_portability/python2.7_sdk]
bb1c706d152d: Preparing
89945359fa87: Preparing
1e511f186fa6: Preparing
1905a350aea9: Preparing
d63ad7b80fd6: Preparing
6e5ed951e507: Preparing
b6052ffcaca8: Preparing
1fcb80ca8dcb: Preparing
7d62a72781b1: Preparing
c28383f3aac4: Preparing
22b7de8c8281: Preparing
6f029f7fa589: Preparing
cbdd44c5cbac: Preparing
2e517d68c391: Preparing
5f3a5adb8e97: Preparing
73bfa217d66f: Preparing
7d62a72781b1: Waiting
6e5ed951e507: Waiting
1fcb80ca8dcb: Waiting
c28383f3aac4: Waiting
cbdd44c5cbac: Waiting
5f3a5adb8e97: Waiting
2e517d68c391: Waiting
91ecdd7165d3: Preparing
e4b20fcc48f4: Preparing
73bfa217d66f: Waiting
91ecdd7165d3: Waiting
d63ad7b80fd6: Pushed
89945359fa87: Pushed
1905a350aea9: Pushed
bb1c706d152d: Pushed
6e5ed951e507: Pushed
c28383f3aac4: Layer already exists
22b7de8c8281: Layer already exists
6f029f7fa589: Layer already exists
cbdd44c5cbac: Layer already exists
2e517d68c391: Layer already exists
5f3a5adb8e97: Layer already exists
73bfa217d66f: Layer already exists
91ecdd7165d3: Layer already exists
e4b20fcc48f4: Layer already exists
1fcb80ca8dcb: Pushed
7d62a72781b1: Pushed
1e511f186fa6: Pushed
b6052ffcaca8: Pushed
latest: digest: sha256:03359efc42866264063f8f53a26c619bf4463b09d8bc85e35a00edf47d2df382 size: 4110
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Pdocker-repository-root=gcr.io/apache-beam-testing/beam_portability -Pdocker-tag=latest :runners:flink:1.9:job-server-container:docker
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:flink:1.9:job-server-container:dockerClean UP-TO-DATE
> Task :runners:flink:1.9:job-server:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :runners:flink:1.9:compileJava FROM-CACHE
> Task :runners:flink:1.9:classes
> Task :runners:flink:1.9:jar
> Task :runners:flink:1.9:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.9:job-server:classes UP-TO-DATE
> Task :runners:flink:1.9:job-server:shadowJar
> Task :runners:flink:1.9:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.9:job-server-container:dockerPrepare
> Task :runners:flink:1.9:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 2s
58 actionable tasks: 40 executed, 17 from cache, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/77hgsnms5s37w

[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins1730302409575363241.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins6621300566872793564.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/flink-job-server gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Error response from daemon: No such image: gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #140

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/140/display/redirect?page=changes>

Changes:

[kcweaver] [BEAM-8139] Test Spark portable job jar.

[kcweaver] [BEAM-8139] Add main method to SparkPipelineRunner.

[kcweaver] [BEAM-8139] Add beam_PostCommit_PortableJar_Spark.

[suztomo] BEAM-8693 com.google.cloud.datastore:datastore-v1-proto-client 1.6.3

[suztomo] google_cloud_bigquery is not declaring google_http_client_jackson

[chadrik] Pass artifact and provision endpoints to external workers from python

[apilloud] [BEAM-8362] Don't use toString() for accessing Enum Types

[echauchot] [BEAM-8894] exclude FlattenWithHeterogeneousCoders category because

[echauchot] [BEAM-8025] Under load we get a NoHostAvailable exception at cluster

[ehudm] [BEAM-3713] pytest migration: py27-cython-pytest

[aaltay] [BEAM-8811] Upgrade Beam pipeline diagrams in docs (#10200)

[bhulette] [BEAM-8933] Update BigQuery proto dependency (#10334)

[pabloem] [BEAM-7746] Add python type hints (part 1) (#9915)

[heejong] [BEAM-8943] SDK harness servers don't shut down properly when SDK

[github] [BEAM-8955] Exclude AvroSchemaTest for Spark Runner (#10358)

[relax] attach values so we don't iterate over everything to verify types

[echauchot] [BEAM-8025] Remove temporary folder rule because it suppresses files on

[echauchot] [BEAM-8025] Disable auto compaction on Cassandra node to avoid race


------------------------------------------
[...truncated 38.16 KB...]
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_fn_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_interactive_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_job_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_job_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_provision_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_provision_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_runner_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_runner_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/endpoints_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/external_transforms_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/metrics_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/schema_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/standard_window_fns_pb2.py>
warning: no files found matching 'README.md'
warning: no files found matching 'NOTICE'
warning: no files found matching 'LICENSE'
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md


> Task :sdks:python:container:py2:copyDockerfileDependencies

> Task :sdks:go:resolveBuildDependencies
Resolving google.golang.org/api: commit='386d4e5f4f92f86e6aec85985761bba4b938a2d5', urls=[https://code.googlesource.com/google-api-go-client]
Resolving google.golang.org/genproto: commit='2b5a72b8730b0b16380010cfe5286c42108d88e7', urls=[https://github.com/google/go-genproto]
Resolving google.golang.org/grpc: commit='7646b5360d049a7ca31e9133315db43456f39e2e', urls=[https://github.com/grpc/grpc-go]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]

> Task :sdks:go:installDependencies
> Task :sdks:go:buildLinuxAmd64
> Task :sdks:go:goBuild

> Task :sdks:python:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/go>

> Task :sdks:python:container:installDependencies
> Task :sdks:python:container:buildDarwinAmd64
> Task :sdks:python:container:buildLinuxAmd64
> Task :sdks:python:container:goBuild
> Task :sdks:python:container:py2:copyLauncherDependencies
> Task :sdks:python:container:py2:dockerPrepare
> Task :sdks:python:container:py2:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 6m 31s
18 actionable tasks: 17 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/d6k43yo5wnvnk

Build step 'Invoke Gradle script' changed build result to SUCCESS
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins3612796172298495341.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins2089895090553814639.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/python2.7_sdk gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins329930964110214245.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins133524380154749947.sh
+ docker push gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
The push refers to repository [gcr.io/apache-beam-testing/beam_portability/python2.7_sdk]
afd6e862e748: Preparing
321c6fc37608: Preparing
382347650233: Preparing
565cb377b2b3: Preparing
e8ddb8a44d33: Preparing
47ab7a8dc5f1: Preparing
ea6116a1f0a8: Preparing
f2815c342f9e: Preparing
9a9ee1e1b077: Preparing
c28383f3aac4: Preparing
22b7de8c8281: Preparing
6f029f7fa589: Preparing
cbdd44c5cbac: Preparing
2e517d68c391: Preparing
5f3a5adb8e97: Preparing
73bfa217d66f: Preparing
91ecdd7165d3: Preparing
e4b20fcc48f4: Preparing
cbdd44c5cbac: Waiting
ea6116a1f0a8: Waiting
f2815c342f9e: Waiting
91ecdd7165d3: Waiting
9a9ee1e1b077: Waiting
5f3a5adb8e97: Waiting
c28383f3aac4: Waiting
e4b20fcc48f4: Waiting
73bfa217d66f: Waiting
47ab7a8dc5f1: Waiting
22b7de8c8281: Waiting
2e517d68c391: Waiting
6f029f7fa589: Waiting
565cb377b2b3: Pushed
321c6fc37608: Pushed
e8ddb8a44d33: Pushed
afd6e862e748: Pushed
f2815c342f9e: Pushed
47ab7a8dc5f1: Pushed
c28383f3aac4: Layer already exists
22b7de8c8281: Layer already exists
6f029f7fa589: Layer already exists
cbdd44c5cbac: Layer already exists
2e517d68c391: Layer already exists
5f3a5adb8e97: Layer already exists
73bfa217d66f: Layer already exists
91ecdd7165d3: Layer already exists
e4b20fcc48f4: Layer already exists
9a9ee1e1b077: Pushed
382347650233: Pushed
ea6116a1f0a8: Pushed
latest: digest: sha256:53b5a65e6a11d087cd57eaff3ca1dfd6d7ff149cd14df4bd25d24bad04582683 size: 4110
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Pdocker-repository-root=gcr.io/apache-beam-testing/beam_portability -Pdocker-tag=latest :runners:flink:1.9:job-server-container:docker
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:flink:1.9:job-server:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:flink:1.9:job-server-container:dockerClean UP-TO-DATE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :runners:flink:1.9:compileJava FROM-CACHE
> Task :runners:flink:1.9:classes
> Task :runners:flink:1.9:jar
> Task :runners:flink:1.9:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.9:job-server:classes UP-TO-DATE
> Task :runners:flink:1.9:job-server:shadowJar
> Task :runners:flink:1.9:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.9:job-server-container:dockerPrepare
> Task :runners:flink:1.9:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 3s
58 actionable tasks: 40 executed, 17 from cache, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/utgnyp76fhnkm

[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins6377015683100352376.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins8928632315234832898.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/flink-job-server gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Error response from daemon: No such image: gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #139

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/139/display/redirect?page=changes>

Changes:

[kirillkozlov] Updated the cost model to favor IO with push-down

[kirillkozlov] BigQueryFilter numSupported method

[kirillkozlov] Update SQL BigQuery doc

[echauchot] [BEAM-8830] flatten: better detect empty collections

[dcavazos] [BEAM-7390] Add code snippet for GroupByKey

[kamil.wasilewski] Strict equality comparision for the version of tensorflow dependency

[kamil.wasilewski] Extract installChicagoTaxiExampleRequirements step

[kamil.wasilewski] Look up log level values using `getattr`

[echauchot] [BEAM-8830] Fix empty dataset creation: use the beam coder wrapper that

[chamikara] [BEAM-8575] Test DoFn context params (#10130)

[github] Add an ML section to python SDK overview (#10233)

[echauchot] [BEAM-8830] Exclude failing SDF tests

[lostluck] [BEAM-8920] Go SDK: faster transforms/filter.Distinct with CombinePerKey


------------------------------------------
[...truncated 36.61 KB...]
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_fn_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_fn_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_interactive_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_job_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_job_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_provision_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_provision_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_runner_api_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/beam_runner_api_pb2_grpc.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/endpoints_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/external_transforms_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/metrics_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/schema_pb2.py>
RefactoringTool: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/portability/api/standard_window_fns_pb2.py>
warning: no files found matching 'README.md'
warning: no files found matching 'NOTICE'
warning: no files found matching 'LICENSE'
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md


> Task :sdks:python:container:py2:copyDockerfileDependencies

> Task :sdks:go:resolveBuildDependencies
Resolving google.golang.org/api: commit='386d4e5f4f92f86e6aec85985761bba4b938a2d5', urls=[https://code.googlesource.com/google-api-go-client]
Resolving google.golang.org/genproto: commit='2b5a72b8730b0b16380010cfe5286c42108d88e7', urls=[https://github.com/google/go-genproto]
Resolving google.golang.org/grpc: commit='7646b5360d049a7ca31e9133315db43456f39e2e', urls=[https://github.com/grpc/grpc-go]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]

> Task :sdks:go:installDependencies
> Task :sdks:go:buildLinuxAmd64
> Task :sdks:go:goBuild

> Task :sdks:python:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/go>

> Task :sdks:python:container:installDependencies
> Task :sdks:python:container:buildDarwinAmd64
> Task :sdks:python:container:buildLinuxAmd64
> Task :sdks:python:container:goBuild
> Task :sdks:python:container:py2:copyLauncherDependencies
> Task :sdks:python:container:py2:dockerPrepare
> Task :sdks:python:container:py2:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 6m 16s
18 actionable tasks: 17 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/o2o56jn2x4xaa

Build step 'Invoke Gradle script' changed build result to SUCCESS
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins1002655630216696545.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins4859558483553138155.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/python2.7_sdk gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins8943762592605815859.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins5392605058302163037.sh
+ docker push gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
The push refers to repository [gcr.io/apache-beam-testing/beam_portability/python2.7_sdk]
50de148c129a: Preparing
770e95d5e88f: Preparing
3de5c6dafaa5: Preparing
d0f50639f84f: Preparing
7aeabf471eb8: Preparing
99dbf750daf6: Preparing
889c76817f27: Preparing
d1e2ef1edbd7: Preparing
9de4c6b04aa8: Preparing
c28383f3aac4: Preparing
22b7de8c8281: Preparing
6f029f7fa589: Preparing
cbdd44c5cbac: Preparing
2e517d68c391: Preparing
5f3a5adb8e97: Preparing
73bfa217d66f: Preparing
91ecdd7165d3: Preparing
e4b20fcc48f4: Preparing
2e517d68c391: Waiting
6f029f7fa589: Waiting
5f3a5adb8e97: Waiting
cbdd44c5cbac: Waiting
d1e2ef1edbd7: Waiting
73bfa217d66f: Waiting
9de4c6b04aa8: Waiting
99dbf750daf6: Waiting
e4b20fcc48f4: Waiting
22b7de8c8281: Waiting
c28383f3aac4: Waiting
889c76817f27: Waiting
d0f50639f84f: Pushed
7aeabf471eb8: Pushed
770e95d5e88f: Pushed
50de148c129a: Pushed
d1e2ef1edbd7: Pushed
c28383f3aac4: Layer already exists
22b7de8c8281: Layer already exists
6f029f7fa589: Layer already exists
99dbf750daf6: Pushed
cbdd44c5cbac: Layer already exists
5f3a5adb8e97: Layer already exists
2e517d68c391: Layer already exists
73bfa217d66f: Layer already exists
e4b20fcc48f4: Layer already exists
91ecdd7165d3: Layer already exists
9de4c6b04aa8: Pushed
3de5c6dafaa5: Pushed
889c76817f27: Pushed
latest: digest: sha256:826b39ed4c7c6330125c831ed4dc121b765046004c7a541a7c8cacaf700bfb1b size: 4110
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Pdocker-repository-root=gcr.io/apache-beam-testing/beam_portability -Pdocker-tag=latest :runners:flink:1.9:job-server-container:docker
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :runners:flink:1.9:job-server:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:flink:1.9:job-server-container:dockerClean UP-TO-DATE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :model:job-management:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :runners:flink:1.9:compileJava FROM-CACHE
> Task :runners:flink:1.9:classes
> Task :runners:flink:1.9:jar
> Task :runners:flink:1.9:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.9:job-server:classes UP-TO-DATE
> Task :runners:flink:1.9:job-server:shadowJar
> Task :runners:flink:1.9:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.9:job-server-container:dockerPrepare
> Task :runners:flink:1.9:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 1s
58 actionable tasks: 40 executed, 17 from cache, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/igcvqoxev7qeu

[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins9103233486093956597.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins4291628246430339323.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/flink-job-server gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Error response from daemon: No such image: gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #138

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/138/display/redirect?page=changes>

Changes:

[dcavazos] [BEAM-7390] Add code snippet for Mean

[nielm] Add limit on number of mutated rows to batching/sorting stages.

[kcweaver] Version Flink job server container images

[kcweaver] [BEAM-8337] publish Flink job server container images

[kcweaver] [BEAM-8870] Skip test_sdf_with_watermark_tracking for Spark

[kcweaver] [BEAM-8870] Cache environment in spark_runner_test.

[heejong] [BEAM-8903] handling --jar_packages experimental flag in PortableRunner

[github] Changing _BatchSizeEstimator defaults to match BatchElements defaults

[ningk] [BEAM-7926] Data-centric Interactive Part1

[kcweaver] Don't override default for environment_cache_millis

[ehudm] [BEAM-3713] pytest migration: py27-gcp-pytest

[sunjincheng121] [BEAM-8929] Remove unnecessary exception handling in

[github] Changing RowAsDictJsonCoder implementation for efficiency (#10300)

[github] Merge pull request #10151: [BEAM-7116] Remove use of KV in Schema

[suztomo] Commons-codec dependency in extensions/sql

[kcweaver] Get Flink version numbers from subdirectories

[suztomo] Adding joda_time dependency

[kcweaver] Warn if Flink versions can't be listed.

[suztomo] Commons_lang3 and jackson_databind dependency

[pabloem] [BEAM-8335] Adds support for multi-output TestStream (#9953)


------------------------------------------
[...truncated 36.29 KB...]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving github.com/spf13/afero: commit='bb8f1927f2a9d3ab41c9340aa034f6b803f4359c', urls=[https://github.com/spf13/afero.git, git@github.com:spf13/afero.git]
Resolving github.com/spf13/cast: commit='acbeb36b902d72a7a4c18e8f3241075e7ab763e4', urls=[https://github.com/spf13/cast.git, git@github.com:spf13/cast.git]
Resolving github.com/spf13/cobra: commit='93959269ad99e80983c9ba742a7e01203a4c0e4f', urls=[https://github.com/spf13/cobra.git, git@github.com:spf13/cobra.git]
Resolving github.com/spf13/jwalterweatherman: commit='7c0cea34c8ece3fbeb2b27ab9b59511d360fb394', urls=[https://github.com/spf13/jwalterweatherman.git, git@github.com:spf13/jwalterweatherman.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving github.com/spf13/viper: commit='aafc9e6bc7b7bb53ddaa75a5ef49a17d6e654be5', urls=[https://github.com/spf13/viper.git, git@github.com:spf13/viper.git]
Resolving github.com/stathat/go: commit='74669b9f388d9d788c97399a0824adbfee78400e', urls=[https://github.com/stathat/go.git, git@github.com:stathat/go.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving github.com/xordataexchange/crypt: commit='b2862e3d0a775f18c7cfe02273500ae307b61218', urls=[https://github.com/xordataexchange/crypt.git, git@github.com:xordataexchange/crypt.git]
Resolving go.opencensus.io: commit='aa2b39d1618ef56ba156f27cfcdae9042f68f0bc', urls=[https://github.com/census-instrumentation/opencensus-go]
Resolving golang.org/x/crypto: commit='d9133f5469342136e669e85192a26056b587f503', urls=[https://go.googlesource.com/crypto]
Resolving golang.org/x/debug: commit='95515998a8a4bd7448134b2cb5971dbeb12e0b77', urls=[https://go.googlesource.com/debug]
Resolving golang.org/x/net: commit='2fb46b16b8dda405028c50f7c7f0f9dd1fa6bfb1', urls=[https://go.googlesource.com/net]
Resolving golang.org/x/oauth2: commit='a032972e28060ca4f5644acffae3dfc268cc09db', urls=[https://go.googlesource.com/oauth2]
Resolving golang.org/x/sync: commit='fd80eb99c8f653c847d294a001bdf2a3a6f768f5', urls=[https://go.googlesource.com/sync]
Resolving golang.org/x/sys: commit='37707fdb30a5b38865cfb95e5aab41707daec7fd', urls=[https://go.googlesource.com/sys]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving google.golang.org/api: commit='386d4e5f4f92f86e6aec85985761bba4b938a2d5', urls=[https://code.googlesource.com/google-api-go-client]
Resolving google.golang.org/genproto: commit='2b5a72b8730b0b16380010cfe5286c42108d88e7', urls=[https://github.com/google/go-genproto]
Resolving google.golang.org/grpc: commit='7646b5360d049a7ca31e9133315db43456f39e2e', urls=[https://github.com/grpc/grpc-go]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]
Resolving cached github.com/coreos/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/coreos/etcd.git, git@github.com:coreos/etcd.git]

> Task :sdks:go:installDependencies
> Task :sdks:go:buildLinuxAmd64
> Task :sdks:go:goBuild

> Task :sdks:python:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/go>

> Task :sdks:python:container:installDependencies
> Task :sdks:python:container:buildDarwinAmd64
> Task :sdks:python:container:buildLinuxAmd64
> Task :sdks:python:container:goBuild
> Task :sdks:python:container:py2:copyLauncherDependencies
> Task :sdks:python:container:py2:dockerPrepare
> Task :sdks:python:container:py2:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 7m 3s
18 actionable tasks: 17 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/7et6hnso2qdiy

Build step 'Invoke Gradle script' changed build result to SUCCESS
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins6552113823138233924.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins5339152940633192187.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/python2.7_sdk gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins1732866002152709877.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins6987951295487820461.sh
+ docker push gcr.io/apache-beam-testing/beam_portability/python2.7_sdk:latest
The push refers to repository [gcr.io/apache-beam-testing/beam_portability/python2.7_sdk]
45555e96037a: Preparing
dc16e2571bb6: Preparing
ccb9c3522f29: Preparing
367aaa275578: Preparing
e619a45abaf4: Preparing
9abbdf8d05df: Preparing
ee5752a42820: Preparing
67f4f5ab49d1: Preparing
758cef2d4df7: Preparing
c28383f3aac4: Preparing
22b7de8c8281: Preparing
6f029f7fa589: Preparing
cbdd44c5cbac: Preparing
2e517d68c391: Preparing
5f3a5adb8e97: Preparing
73bfa217d66f: Preparing
91ecdd7165d3: Preparing
e4b20fcc48f4: Preparing
758cef2d4df7: Waiting
c28383f3aac4: Waiting
9abbdf8d05df: Waiting
ee5752a42820: Waiting
67f4f5ab49d1: Waiting
cbdd44c5cbac: Waiting
2e517d68c391: Waiting
22b7de8c8281: Waiting
5f3a5adb8e97: Waiting
6f029f7fa589: Waiting
73bfa217d66f: Waiting
91ecdd7165d3: Waiting
e4b20fcc48f4: Waiting
dc16e2571bb6: Pushed
e619a45abaf4: Pushed
367aaa275578: Pushed
45555e96037a: Pushed
9abbdf8d05df: Pushed
c28383f3aac4: Layer already exists
22b7de8c8281: Layer already exists
6f029f7fa589: Layer already exists
cbdd44c5cbac: Layer already exists
67f4f5ab49d1: Pushed
2e517d68c391: Layer already exists
73bfa217d66f: Layer already exists
5f3a5adb8e97: Layer already exists
e4b20fcc48f4: Layer already exists
91ecdd7165d3: Layer already exists
758cef2d4df7: Pushed
ccb9c3522f29: Pushed
ee5752a42820: Pushed
latest: digest: sha256:a39dde79c52a508929ebdf631c68608c81b1c4bffc3a6f7c3e74f2b877db466c size: 4110
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Pdocker-repository-root=gcr.io/apache-beam-testing/beam_portability -Pdocker-tag=latest :runners:flink:1.9:job-server-container:docker
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:flink:1.9:job-server:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:protobuf:extractProto
> Task :runners:flink:1.9:job-server-container:dockerClean UP-TO-DATE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:build-tools:jar
> Task :sdks:java:core:processResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:io:kafka:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :runners:flink:1.9:compileJava FROM-CACHE
> Task :runners:flink:1.9:classes
> Task :runners:flink:1.9:jar
> Task :runners:flink:1.9:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.9:job-server:classes UP-TO-DATE
> Task :runners:flink:1.9:job-server:shadowJar
> Task :runners:flink:1.9:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.9:job-server-container:dockerPrepare
> Task :runners:flink:1.9:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 4s
58 actionable tasks: 40 executed, 17 from cache, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/y4md2ircdyelk

[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins4340480797708791926.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_coGBK_Flink_Batch] $ /bin/bash -xe /tmp/jenkins7771749885700078627.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/flink-job-server gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Error response from daemon: No such image: gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #137

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/137/display/redirect?page=changes>

Changes:

[michal.walenia] [BEAM-8895] Add BigQuery table name sanitization to BigQueryIOIT

[michal.walenia] [BEAM-8918] Split batch BQIOIT into avro and json using tests


------------------------------------------
[...truncated 215.40 KB...]
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:flink_version:v1"
    value {
      string_value: "1.9"
    }
  }
  fields {
    key: "beam:option:gcs_performance_metrics:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:job_endpoint:v1"
    value {
      string_value: "localhost:8099"
    }
  }
  fields {
    key: "beam:option:job_name:v1"
    value {
      string_value: "load-tests-python-flink-batch-cogbk-3-1209150527"
    }
  }
  fields {
    key: "beam:option:job_port:v1"
    value {
      string_value: "0"
    }
  }
  fields {
    key: "beam:option:job_server_timeout:v1"
    value {
      string_value: "60"
    }
  }
  fields {
    key: "beam:option:no_auth:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:object_reuse:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:parallelism:v1"
    value {
      string_value: "5"
    }
  }
  fields {
    key: "beam:option:pipeline_type_check:v1"
    value {
      bool_value: true
    }
  }
  fields {
    key: "beam:option:profile_cpu:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_memory:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_sample_rate:v1"
    value {
      number_value: 1.0
    }
  }
  fields {
    key: "beam:option:project:v1"
    value {
      string_value: "apache-beam-testing"
    }
  }
  fields {
    key: "beam:option:retain_docker_containers:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:retain_externalized_checkpoints_on_cancellation:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:runtime_type_check:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:save_main_session:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:sdk_location:v1"
    value {
      string_value: "container"
    }
  }
  fields {
    key: "beam:option:sdk_worker_parallelism:v1"
    value {
      string_value: "1"
    }
  }
  fields {
    key: "beam:option:shutdown_sources_on_final_watermark:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:streaming:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:temp_location:v1"
    value {
      string_value: "gs://temp-storage-for-perf-tests/loadtests"
    }
  }
  fields {
    key: "beam:option:type_check_strictness:v1"
    value {
      string_value: "DEFAULT_TO_ANY"
    }
  }
  fields {
    key: "beam:option:update:v1"
    value {
      bool_value: false
    }
  }
}
job_name: "job"

apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: org.apache.flink.client.program.ProgramInvocationException: Job failed. (JobID: e45a874cec6455db7ea273ff16cdc53f)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:301)
	at org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:209)
	at org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:186)
	at org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:173)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:191)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:116)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:84)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:81)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
	... 16 more
Caused by: java.util.concurrent.TimeoutException: Heartbeat of TaskManager with id container_e01_1575908212299_0001_01_000003 timed out.
	at org.apache.flink.runtime.jobmaster.JobMaster$TaskManagerHeartbeatListener.notifyHeartbeatTimeout(JobMaster.java:1149)
	at org.apache.flink.runtime.heartbeat.HeartbeatMonitorImpl.run(HeartbeatMonitorImpl.java:109)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:397)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:190)
	at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
	at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
	at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
	at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
	at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
	at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
	at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
	at akka.actor.ActorCell.invoke(ActorCell.scala:561)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
	at akka.dispatch.Mailbox.run(Mailbox.scala:225)
	at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
	at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

root: ERROR: java.util.concurrent.TimeoutException: Heartbeat of TaskManager with id container_e01_1575908212299_0001_01_000003 timed out.
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 1676.651s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 55

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 28m 1s
3 actionable tasks: 2 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/ma7ml5dkuspi4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_coGBK_Flink_Batch #136

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/136/display/redirect>

Changes:


------------------------------------------
[...truncated 244.62 KB...]
      string_value: "1.9"
    }
  }
  fields {
    key: "beam:option:gcs_performance_metrics:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:job_endpoint:v1"
    value {
      string_value: "localhost:8099"
    }
  }
  fields {
    key: "beam:option:job_name:v1"
    value {
      string_value: "load-tests-python-flink-batch-cogbk-4-1208150525"
    }
  }
  fields {
    key: "beam:option:job_port:v1"
    value {
      string_value: "0"
    }
  }
  fields {
    key: "beam:option:job_server_timeout:v1"
    value {
      string_value: "60"
    }
  }
  fields {
    key: "beam:option:no_auth:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:object_reuse:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:parallelism:v1"
    value {
      string_value: "5"
    }
  }
  fields {
    key: "beam:option:pipeline_type_check:v1"
    value {
      bool_value: true
    }
  }
  fields {
    key: "beam:option:profile_cpu:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_memory:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:profile_sample_rate:v1"
    value {
      number_value: 1.0
    }
  }
  fields {
    key: "beam:option:project:v1"
    value {
      string_value: "apache-beam-testing"
    }
  }
  fields {
    key: "beam:option:retain_docker_containers:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:retain_externalized_checkpoints_on_cancellation:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:runtime_type_check:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:save_main_session:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:sdk_location:v1"
    value {
      string_value: "container"
    }
  }
  fields {
    key: "beam:option:sdk_worker_parallelism:v1"
    value {
      string_value: "1"
    }
  }
  fields {
    key: "beam:option:shutdown_sources_on_final_watermark:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:streaming:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:temp_location:v1"
    value {
      string_value: "gs://temp-storage-for-perf-tests/loadtests"
    }
  }
  fields {
    key: "beam:option:type_check_strictness:v1"
    value {
      string_value: "DEFAULT_TO_ANY"
    }
  }
  fields {
    key: "beam:option:update:v1"
    value {
      bool_value: false
    }
  }
}
job_name: "job"

apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: org.apache.flink.client.program.ProgramInvocationException: Job failed. (JobID: 909d4608f30d38cf92bfca71032da91c)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
	at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:301)
	at org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:209)
	at org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:186)
	at org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:173)
	at org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:191)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:116)
	at org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:84)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:81)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
	at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
	at org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
	... 16 more
Caused by: java.lang.Exception: The data preparation for task 'GroupReduce (GroupReduce at CoGroupByKey /GroupByKey)' , caused an error: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-136-w-1.c.apache-beam-testing.internal/10.128.0.118:38785'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480)
	at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
	at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
	at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
	... 1 more
Caused by: java.lang.RuntimeException: Error obtaining the sorted input: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-136-w-1.c.apache-beam-testing.internal/10.128.0.118:38785'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger.getIterator(UnilateralSortMerger.java:650)
	at org.apache.flink.runtime.operators.BatchTask.getInput(BatchTask.java:1109)
	at org.apache.flink.runtime.operators.GroupReduceDriver.prepare(GroupReduceDriver.java:99)
	at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:474)
	... 4 more
Caused by: java.io.IOException: Thread 'SortMerger Reading Thread' terminated due to an exception: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-136-w-1.c.apache-beam-testing.internal/10.128.0.118:38785'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.operators.sort.UnilateralSortMerger$ThreadBase.run(UnilateralSortMerger.java:831)
Caused by: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-136-w-1.c.apache-beam-testing.internal/10.128.0.118:38785'. This might indicate that the remote task manager was lost.
	at org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.channelInactive(CreditBasedPartitionRequestClientHandler.java:136)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
	at org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInputClosed(ByteToMessageDecoder.java:390)
	at org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInactive(ByteToMessageDecoder.java:355)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1429)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
	at org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:947)
	at org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:826)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:404)
	at org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:474)
	at org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:909)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connection unexpectedly closed by remote task manager 'beam-loadtests-python-cogbk-flink-batch-136-w-1.c.apache-beam-testing.internal/10.128.0.118:38785'. This might indicate that the remote task manager was lost.
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 947.638s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_coGBK_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 55

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 52s
3 actionable tasks: 2 executed, 1 up-to-date

Publishing build scan...
https://scans.gradle.com/s/fj6js6dbbgkzy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org