You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/12/17 18:12:43 UTC

Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9731

See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9731/display/redirect>

Changes:


------------------------------------------
[...truncated 15.24 KB...]
> Task :model:fn-execution:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :examples:java:processTestResources
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :model:job-management:processResources
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :runners:spark:2:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :model:job-management:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :examples:java:twitter:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:testJar
> Task :runners:spark:3:testJar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 12m 8s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/zpbxjescyb33c

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Java_ValidatesRunner_Spark #9811

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9811/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9810

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9810/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 90f167561b34725693649c1485dacc299adbcf16 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 90f167561b34725693649c1485dacc299adbcf16 # timeout=10
Commit message: "[BEAM-10345] Add an import guard to support recent google-cloud-spanner versions. (#16434)"
 > git rev-list --no-walk 90f167561b34725693649c1485dacc299adbcf16 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 9s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/c7tsmoxbb6edm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9809

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9809/display/redirect?page=changes>

Changes:

[ningkang0957] [BEAM-12879] Prevented missing permission from failing GCS I/O

[noreply] Merge pull request #16347: fix: move connector to use v1 BigQuery

[noreply] [BEAM-13603] Fix bug in apache_beam.utils.Shared (#16437)

[noreply] [BEAM-10345] Add an import guard to support recent google-cloud-spanner


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-6 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 90f167561b34725693649c1485dacc299adbcf16 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 90f167561b34725693649c1485dacc299adbcf16 # timeout=10
Commit message: "[BEAM-10345] Add an import guard to support recent google-cloud-spanner versions. (#16434)"
 > git rev-list --no-walk 46b3ef598b924e46d519c72e7eca211b4efffe82 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 25s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/zsy7fyc2lfat2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9808

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9808/display/redirect?page=changes>

Changes:

[aydar.zaynutdinov] [BEAM-13565][Playground]

[relax] update bom

[zyichi] Upgrade python library versions in base_image_requirements.txt

[noreply] [BEAM-13601] Don't cache Row types for a schema. (#16427)

[noreply] [BEAM-13430] Re-enable checkerframework (#16429)

[noreply] [BEAM-13430] Ensure that testRuntimeMigration depends on "default"

[noreply] Merge pull request #16277 from [BEAM-13124][Playground] Create readiness

[noreply] Merge pull request #16314 from [BEAM-13260][Playground]Implement setup

[noreply] Merge pull request #16383 from [BEAM-13566][Playground] Add logic of

[noreply] Merge pull request #16365 from [BEAM-13559][Playground] Remove tag in

[noreply] Merge pull request #16360 from [BEAM-13546][Playground] Update nginx

[noreply] Merge pull request #16192 from [BEAM-13395] [Playground] Tag katas

[noreply] Merge pull request #16254 from [BEAM-13249][Playground] Security – Mock


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 46b3ef598b924e46d519c72e7eca211b4efffe82 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 46b3ef598b924e46d519c72e7eca211b4efffe82 # timeout=10
Commit message: "Merge pull request #16377: [BEAM-13576] update bom"
 > git rev-list --no-walk 38b833aa1dd0685be9137f41f6071c55f0c50b46 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon (subsequent builds will be faster)
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 42s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/qnqczhf2r6tvm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9807

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9807/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13567] Consolidate runner flag definition. (#16426)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 38b833aa1dd0685be9137f41f6071c55f0c50b46 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 38b833aa1dd0685be9137f41f6071c55f0c50b46 # timeout=10
Commit message: "[BEAM-13567] Consolidate runner flag definition. (#16426)"
 > git rev-list --no-walk e9e49ff1fcb516350de7ad6affc80455373d7f38 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 9s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/yzv434kwffvxg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9806

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9806/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision e9e49ff1fcb516350de7ad6affc80455373d7f38 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f e9e49ff1fcb516350de7ad6affc80455373d7f38 # timeout=10
Commit message: "[BEAM-13015] Use 20% of memory when the maximum has been configured. (#16420)"
 > git rev-list --no-walk e9e49ff1fcb516350de7ad6affc80455373d7f38 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 7s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ahzcubhhyzdzk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9805

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9805/display/redirect?page=changes>

Changes:

[Steve Niemitz] [BEAM-13459] Update CHANGES.md, add note about artifact caching for

[zyichi] Fix remaining failing perf IT tests.

[noreply] [BEAM-13015] Use 20% of memory when the maximum has been configured.


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision e9e49ff1fcb516350de7ad6affc80455373d7f38 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f e9e49ff1fcb516350de7ad6affc80455373d7f38 # timeout=10
Commit message: "[BEAM-13015] Use 20% of memory when the maximum has been configured. (#16420)"
 > git rev-list --no-walk ea65a054f2fcb6349478d19609a773f66bbfa20e # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 57s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/3uvg5isya7sbm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9804

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9804/display/redirect?page=changes>

Changes:

[johnjcasey] [BEAM-12391] update avro sink to close the opened file handle, instead

[Robert Bradshaw] [BEAM-13482] Python fully qualified name external transforms.

[Robert Bradshaw] Add Python expansion service entry point.

[Kyle Weaver] [BEAM-13569] Change Spark dependencies to implementation.

[Kyle Weaver] remove redundant dependency

[wuren] [BEAM-13591] Bump log4j2 version to 2.17.1

[relax] Add Flink runner support for OrderedListState. This version reads the

[zyichi] Fix sdk_container_builder too many values to unpack error

[noreply] [BEAM-13480] Sickbay PubSubIntegrationTest.test_streaming_data_only on

[Kyle Weaver] remove redundant testImplementation dependencies

[noreply] [BEAM-13430] Swap to use "mainClass" instead of "main" since it was

[noreply] [BEAM-13430] Replace deprecated "appendix" with "archiveAppendix"

[noreply] [BEAM-13015] Add jamm as a java agent to the Java SDK harness container

[noreply] [BEAM-13430] Partially revert

[noreply] Merge pull request #15863 from [BEAM-13184] Autosharding for

[noreply] [BEAM-11936] Enable FloatingPointAssertionWithinEpsilon errorprone check

[noreply] [BEAM-11936] Enable LockNotBeforeTry errorprone check (#16259)

[noreply] [BEAM-11936] Enable errorprone unused checks (#16262)

[noreply] Add Nexmark Query 14 (#16337)

[noreply] [BEAM-13015] Migrate all user state and side implementations to support


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-11 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision ea65a054f2fcb6349478d19609a773f66bbfa20e (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f ea65a054f2fcb6349478d19609a773f66bbfa20e # timeout=10
Commit message: "[BEAM-13015] Migrate all user state and side implementations to support caching. (#16263)"
 > git rev-list --no-walk 17b62ad9e050f80a88793457aee710ea4711d47b # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 59s
10 actionable tasks: 6 executed, 3 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/neid6ywnnz6x2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9803

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9803/display/redirect?page=changes>

Changes:

[noreply] Merge pull request #16404: [BEAM-13586] Fix NPE in


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 17b62ad9e050f80a88793457aee710ea4711d47b (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 17b62ad9e050f80a88793457aee710ea4711d47b # timeout=10
Commit message: "Merge pull request #16404: [BEAM-13586] Fix NPE in DefaultS3ClientBuilderFactory"
 > git rev-list --no-walk c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 3s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/acgqxqaol7a6c

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9802

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9802/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision c4fc3987f31a57a31d2081297fa639084b35d391 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
Commit message: "Merge pull request #16368: [BEAM-13509] Stop sharing SQS client between readers of same source."
 > git rev-list --no-walk c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/enojttzkzcxmk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9801

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9801/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision c4fc3987f31a57a31d2081297fa639084b35d391 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
Commit message: "Merge pull request #16368: [BEAM-13509] Stop sharing SQS client between readers of same source."
 > git rev-list --no-walk c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon (subsequent builds will be faster)
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 25s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/l6whxm3qghf7m

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9800

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9800/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision c4fc3987f31a57a31d2081297fa639084b35d391 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
Commit message: "Merge pull request #16368: [BEAM-13509] Stop sharing SQS client between readers of same source."
 > git rev-list --no-walk c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 56s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/oy22yr7p2ik3w

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9799

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9799/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13509] Stop sharing SQS client between readers of same source.

[mmack] [BEAM-13587] Attempt to load AWS region from default provider chain in


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision c4fc3987f31a57a31d2081297fa639084b35d391 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f c4fc3987f31a57a31d2081297fa639084b35d391 # timeout=10
Commit message: "Merge pull request #16368: [BEAM-13509] Stop sharing SQS client between readers of same source."
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 8s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/s5dw6scyu7qfu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9798

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9798/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 2s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ufzf3ncyrs2ly

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9797

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9797/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon (subsequent builds will be faster)
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 36s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ajvig2lbqcgbs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9796

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9796/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 43s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/fj7qaxwipdxfy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9795

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9795/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 52s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/4tvjrei4lofy2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9794

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9794/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-4 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 3 busy and 1 incompatible Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 20s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/2ax2x4j2qasr6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9793

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9793/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 13s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/4vosbxy62kbhm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9792

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9792/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 7487bf0118e4c7704793112f904c7cf328caebfe (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 7487bf0118e4c7704793112f904c7cf328caebfe # timeout=10
Commit message: "[BEAM-13498, BEAM-13522] Sickbay broken test and test category for Spark (#16411)"
 > git rev-list --no-walk 34118c690783b8e0706932ca621b26b316476c2f # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 59s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/drxnnrscmij3a

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9791

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9791/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 34118c690783b8e0706932ca621b26b316476c2f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 34118c690783b8e0706932ca621b26b316476c2f # timeout=10
Commit message: "[BEAM-12092] Bump jedis to version 3.8.0 (#16403)"
 > git rev-list --no-walk 34118c690783b8e0706932ca621b26b316476c2f # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 55s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/yu7oa2rd66wk4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9790

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9790/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 34118c690783b8e0706932ca621b26b316476c2f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 34118c690783b8e0706932ca621b26b316476c2f # timeout=10
Commit message: "[BEAM-12092] Bump jedis to version 3.8.0 (#16403)"
 > git rev-list --no-walk 34118c690783b8e0706932ca621b26b316476c2f # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 56s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ellhhg5bkrqao

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9789

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9789/display/redirect?page=changes>

Changes:

[noreply] [BEAM-12092] Bump jedis to version 3.8.0 (#16403)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 34118c690783b8e0706932ca621b26b316476c2f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 34118c690783b8e0706932ca621b26b316476c2f # timeout=10
Commit message: "[BEAM-12092] Bump jedis to version 3.8.0 (#16403)"
 > git rev-list --no-walk fdbb90b144cf4b037860859bdc0933ebedf05b69 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 1s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/jpr44ekfu2sy6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9788

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9788/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13541] Fix spotbugs issue with DLS in tail (#16407)

[noreply] [BEAM-13588] Add missing module for PVR tests and also sickbay known

[noreply] [BEAM-13575] Sickbay test that is flaky to restore precommit test signal


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision fdbb90b144cf4b037860859bdc0933ebedf05b69 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f fdbb90b144cf4b037860859bdc0933ebedf05b69 # timeout=10
Commit message: "[BEAM-13575] Sickbay test that is flaky to restore precommit test signal (#16410)"
 > git rev-list --no-walk 304680d4a89c7c1797c50e0bd68ea5bedcff68fa # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 33s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ijufucvox5jxg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9787

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9787/display/redirect?page=changes>

Changes:

[stranniknm] [BEAM-13555] playground - quick fix for safari

[noreply] Merge pull request #15769 from [BEAM-13031] [Playground] Code editor -


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-12 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 304680d4a89c7c1797c50e0bd68ea5bedcff68fa (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 304680d4a89c7c1797c50e0bd68ea5bedcff68fa # timeout=10
Commit message: "Merge pull request #16362 from [BEAM-13555] [Bugfix] [Playground] change html renderer for safari to fix freezes"
 > git rev-list --no-walk 3ab26ce26711ea285eafcfe1f01ceed24912e62c # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 40s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/j4sgqx35k4xtu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9786

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9786/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 3ab26ce26711ea285eafcfe1f01ceed24912e62c (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 3ab26ce26711ea285eafcfe1f01ceed24912e62c # timeout=10
Commit message: "[BEAM-13541] More intelligent caching of CoGBK values. (#16354)"
 > git rev-list --no-walk 3ab26ce26711ea285eafcfe1f01ceed24912e62c # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/zo7tf3auscx66

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9785

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9785/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13484][Playground] Improve Terraform Scripts For Deploy

[noreply] [BEAM-13541] More intelligent caching of CoGBK values. (#16354)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 3ab26ce26711ea285eafcfe1f01ceed24912e62c (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 3ab26ce26711ea285eafcfe1f01ceed24912e62c # timeout=10
Commit message: "[BEAM-13541] More intelligent caching of CoGBK values. (#16354)"
 > git rev-list --no-walk 2a9e8b698fd75eb788e4e12f08d3a337f0ff1727 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy and 1 stopped Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 42s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ta2c6nxz4nmuw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9784

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9784/display/redirect?page=changes>

Changes:

[zyichi] Fix class not found in perf integration tests.

[noreply] Update

[noreply] [BEAM-13430] Fix class not found error for example integration tests and

[noreply] [BEAM-13430] Re-enable dependency analysis for modules. (#16395)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-15 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 2a9e8b698fd75eb788e4e12f08d3a337f0ff1727 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 2a9e8b698fd75eb788e4e12f08d3a337f0ff1727 # timeout=10
Commit message: "Merge pull request #16398 from y1chi/perf-it"
 > git rev-list --no-walk fe456b79419d1a67ebf13d7d4b6695fa1aa6204d # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 20s
10 actionable tasks: 5 executed, 3 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/e5kfs7xluk2rs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9783

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9783/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13009] Fix tests to ensure absence of duplicates only per request

[mmack] [adhoc] Use predefined min/max combiners in SQS reader to gather stats


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision fe456b79419d1a67ebf13d7d4b6695fa1aa6204d (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f fe456b79419d1a67ebf13d7d4b6695fa1aa6204d # timeout=10
Commit message: "Merge pull request #16367: [BEAM-13009] Fix tests to ensure absence of duplicates only per request"
 > git rev-list --no-walk 0af368995720dd3723f7d3373982a7105a5d8078 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 6s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/eas6tmoixbica

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9782

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9782/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 0af368995720dd3723f7d3373982a7105a5d8078 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 0af368995720dd3723f7d3373982a7105a5d8078 # timeout=10
Commit message: "[BEAM-13430] Re-enable checkerframework for the project excluding the two classes it started to crash on. (#16389)"
 > git rev-list --no-walk 0af368995720dd3723f7d3373982a7105a5d8078 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 48s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/d2woowyfh4f2s

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9781

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9781/display/redirect?page=changes>

Changes:

[Steve Niemitz] [BEAM-13488] Work around bug in classpath detection

[Steve Niemitz] [BEAM-13459] Cache artifacts across job runs on python+dataflow

[Kyle Weaver] [BEAM-13581] Remove previous job name for Flink PVR precommit.

[emilyye] Moving to 2.37.0-SNAPSHOT on master branch.

[noreply] Release website update for Beam 2.35.0 (#16115)

[noreply] [BEAM-13430] Re-enable checkerframework for the project excluding the


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 0af368995720dd3723f7d3373982a7105a5d8078 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 0af368995720dd3723f7d3373982a7105a5d8078 # timeout=10
Commit message: "[BEAM-13430] Re-enable checkerframework for the project excluding the two classes it started to crash on. (#16389)"
 > git rev-list --no-walk 8eaa0aa33214e7ea23c39e0023bea4ae28bd84de # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 5s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/2s3ot5eabfkcs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9780

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9780/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13578] Validate that repeatable PipelineOptions need to be

[noreply] [BEAM-13430] Update comment over eclipse plugin. (#16387)

[zyichi] [BEAM-13430] Fix broken ULR validates runner tests

[noreply] Update venv creation commands in release scripts. (#16381)

[noreply] [BEAM-12777] Stable URL for current API docs (#16327)

[noreply] [BEAM-13539] correct resource override dependencies (#16392)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-15 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 8eaa0aa33214e7ea23c39e0023bea4ae28bd84de (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 8eaa0aa33214e7ea23c39e0023bea4ae28bd84de # timeout=10
Commit message: "[BEAM-13539] correct resource override dependencies (#16392)"
 > git rev-list --no-walk b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 6s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/hpi5xbgv53a5q

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9779

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9779/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-6 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision b33bebcda54749bc93cc18cb6c3ea03862656b13 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
Commit message: "Merge pull request #16215 from dpcollins-google/publish-no-desync"
 > git rev-list --no-walk b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 24s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ka6raonkv6wyw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9778

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9778/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision b33bebcda54749bc93cc18cb6c3ea03862656b13 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
Commit message: "Merge pull request #16215 from dpcollins-google/publish-no-desync"
 > git rev-list --no-walk b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 52s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/tjhmhmgyj3b5m

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9777

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9777/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision b33bebcda54749bc93cc18cb6c3ea03862656b13 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
Commit message: "Merge pull request #16215 from dpcollins-google/publish-no-desync"
 > git rev-list --no-walk b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/24huqqphi3j7g

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9776

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9776/display/redirect?page=changes>

Changes:

[dpcollins] [BEAM-13402] Simplify PubsubLiteSink

[Kyle Weaver] [BEAM-13571] Fix ClassNotFound exception in Flink tests

[Kyle Weaver] [BEAM-13498] [BEAM-13573] exclude new tests on Flink

[noreply] Exclude UsesOnWindowExpiration by category from Dataflow v2 streaming

[noreply] [BEAM-13052] Increment pubsub python version and fix breakages. (#16126)

[noreply] [BEAM-13052] Add Pub/Sub Lite xlang transforms in python (#15727)

[noreply] [BEAM-13402] Version bump Pub/Sub Lite and implement changes to ensure


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision b33bebcda54749bc93cc18cb6c3ea03862656b13 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f b33bebcda54749bc93cc18cb6c3ea03862656b13 # timeout=10
Commit message: "Merge pull request #16215 from dpcollins-google/publish-no-desync"
 > git rev-list --no-walk 5b06fa3fff0ab784680f10ae500aa6582c9afeec # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 59s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/e5togfwnoasra

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9775

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9775/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-13521] flink pvr batch precommit


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 5b06fa3fff0ab784680f10ae500aa6582c9afeec (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 5b06fa3fff0ab784680f10ae500aa6582c9afeec # timeout=10
Commit message: "Merge pull request #16349 from ibzib/flink-pvr"
 > git rev-list --no-walk f8b5d01077de9eee19a033b82494f7168c43ae94 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 57s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/cumky6n265huu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9774

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9774/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13526] Kafka.IO: make DeserializerProvider a public interface


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision f8b5d01077de9eee19a033b82494f7168c43ae94 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f f8b5d01077de9eee19a033b82494f7168c43ae94 # timeout=10
Commit message: "[BEAM-13526] Kafka.IO: make DeserializerProvider a public interface (#16339)"
 > git rev-list --no-walk 6e735c1e9b8a69cb97c659478966f2c3162c3b50 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/vfuhxjn6ck3da

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9773

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9773/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6e735c1e9b8a69cb97c659478966f2c3162c3b50 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6e735c1e9b8a69cb97c659478966f2c3162c3b50 # timeout=10
Commit message: "Merge pull request #16341: [adhoc] Forbid to import guava and others from org.testcontainers.shaded"
 > git rev-list --no-walk 6e735c1e9b8a69cb97c659478966f2c3162c3b50 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon (subsequent builds will be faster)
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 33s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ls6uhmenscu6i

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9772

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9772/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6e735c1e9b8a69cb97c659478966f2c3162c3b50 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6e735c1e9b8a69cb97c659478966f2c3162c3b50 # timeout=10
Commit message: "Merge pull request #16341: [adhoc] Forbid to import guava and others from org.testcontainers.shaded"
 > git rev-list --no-walk 6e735c1e9b8a69cb97c659478966f2c3162c3b50 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon (subsequent builds will be faster)
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 23s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/lfxjvwumty3ea

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9771

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9771/display/redirect?page=changes>

Changes:

[mmack] [adhoc] Forbid to import guava and others from org.testcontainers.shaded


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6e735c1e9b8a69cb97c659478966f2c3162c3b50 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6e735c1e9b8a69cb97c659478966f2c3162c3b50 # timeout=10
Commit message: "Merge pull request #16341: [adhoc] Forbid to import guava and others from org.testcontainers.shaded"
 > git rev-list --no-walk 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 5s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/d7u2rzx4e5544

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9770

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9770/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
Commit message: "Merge pull request #16285 from [BEAM-13492][Playground]  Update backend README file"
 > git rev-list --no-walk 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 13s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/prgrtttmu27du

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9769

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9769/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
Commit message: "Merge pull request #16285 from [BEAM-13492][Playground]  Update backend README file"
 > git rev-list --no-walk 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 49s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/vjfoann2wt23k

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9768

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9768/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
Commit message: "Merge pull request #16285 from [BEAM-13492][Playground]  Update backend README file"
 > git rev-list --no-walk 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 7s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/eynr6lwx3erc4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9767

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9767/display/redirect?page=changes>

Changes:

[noreply] Merge pull request #16285 from [BEAM-13492][Playground]  Update backend


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6d35bf9311c3b07ffda048de5da8ffe8e6bbcfe6 # timeout=10
Commit message: "Merge pull request #16285 from [BEAM-13492][Playground]  Update backend README file"
 > git rev-list --no-walk 1fa1e3cc7aedda33d208115950309cf6278eef4d # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 18s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/pny4augzdepqe

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9766

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9766/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 1fa1e3cc7aedda33d208115950309cf6278eef4d (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 1fa1e3cc7aedda33d208115950309cf6278eef4d # timeout=10
Commit message: "Merge pull request #16356 from [BEAM-13530] [Playground] [Bugfix] Fix Scrollbar in the Output section"
 > git rev-list --no-walk 1fa1e3cc7aedda33d208115950309cf6278eef4d # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 10s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/miemi2mrvz5uw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9765

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9765/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 1fa1e3cc7aedda33d208115950309cf6278eef4d (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 1fa1e3cc7aedda33d208115950309cf6278eef4d # timeout=10
Commit message: "Merge pull request #16356 from [BEAM-13530] [Playground] [Bugfix] Fix Scrollbar in the Output section"
 > git rev-list --no-walk 1fa1e3cc7aedda33d208115950309cf6278eef4d # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 44s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/7ilpts236rbro

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9764

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9764/display/redirect?page=changes>

Changes:

[alexander.zhuravlev] [BEAM-13530] Fixed scrollbar in the Output section

[noreply] Merge pull request #16355 from [BEAM-13529] [Playground] [Bugfix] Fix

[noreply] Fixed grammar error in Go ParDo function comments. (#16329)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-11 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 1fa1e3cc7aedda33d208115950309cf6278eef4d (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 1fa1e3cc7aedda33d208115950309cf6278eef4d # timeout=10
Commit message: "Merge pull request #16356 from [BEAM-13530] [Playground] [Bugfix] Fix Scrollbar in the Output section"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 55s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/p57yveemyxx7k

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9763

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9763/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 48s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/t2cgj2nbf325a

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9762

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9762/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-13 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 49s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/6tdeumnjiyira

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9761

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9761/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon (subsequent builds will be faster)
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 21s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/rc3k2tclwicqi

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9760

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9760/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 12s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/igcun5dys7y4y

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9759

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9759/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 1s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/jyhffaupr3ckw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9758

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9758/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 58s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ijeor5kpbq3hw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9757

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9757/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13490] Bump log4j version to 2.17.0 (#16276)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 206e3da988d02e7f63b4d91d8df224f8bbb40354 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 206e3da988d02e7f63b4d91d8df224f8bbb40354 # timeout=10
Commit message: "[BEAM-13490] Bump log4j version to 2.17.0 (#16276)"
 > git rev-list --no-walk 37d5ba159543060ebb81fe82303cf849337b7a85 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerBatch'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 17s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/iarwqejlizin4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9756

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9756/display/redirect?page=changes>

Changes:

[alexander.zhuravlev] [BEAM-13489] Updated manifest.json and index.html to follow best

[stranniknm] [BEAM-13531] playground - fetch example info only once

[noreply] Merge pull request #16320 from [BEAM-13512] [Playground] send request to

[noreply] Merge pull request #16209 from [Playground][BEAM-13304]Deploy Python

[noreply] Merge pull request #16105 from [BEAM-13328] [Playground] Add

[noreply] Merge pull request #16103 from [BEAM-13252] [Playground] Feedback Thumbs

[noreply] Better type inference for GroupBy. (#16318)

[noreply] [BEAM-13534] Update TestAutomatedExpansionService liveness check to not


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-11 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 37d5ba159543060ebb81fe82303cf849337b7a85 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 37d5ba159543060ebb81fe82303cf849337b7a85 # timeout=10
Commit message: "[BEAM-13534] Update TestAutomatedExpansionService liveness check to not depend on exec call (#16350)"
 > git rev-list --no-walk 4774ac713f427fefb38114f661516faef26d8207 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 19s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/pmrh6pi4xr3ti

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9755

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9755/display/redirect?page=changes>

Changes:

[mmack] [BEAM-7559] Add integration test for DynamoDBIO (SDK v2)

[mmack] [BEAM-7559] Add integration test for DynamoDBIO (SDK v1)

[noreply] [BEAM-13399] Disable expansion service integration test on Samza and

[noreply] [BEAM-13430] Upgrade beam to gradle 7.3.2 (#16319)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-3 (beam) in workspace <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 4774ac713f427fefb38114f661516faef26d8207 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 4774ac713f427fefb38114f661516faef26d8207 # timeout=10
Commit message: "[BEAM-13430] Upgrade beam to gradle 7.3.2 (#16319)"
 > git rev-list --no-walk c9e81042998da6b05e587710886a8ed54d233b7a # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :runners:spark:2:validatesRunner :runners:spark:3:validatesRunner
Starting a Gradle Daemon, 2 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task ':runners:spark:2:validatesRunnerStreaming'.
> Could not resolve all task dependencies for configuration ':runners:spark:2:validatesRunner'.
   > Could not resolve project :runners:spark:2.
     Required by:
         project :runners:spark:2
      > Selected configuration 'compileClasspath' on 'project :runners:spark:2' but it can't be used as a project dependency because it isn't intended for consumption by other components.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 12s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/ovuhuy6nr2kpc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9754

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9754/display/redirect?page=changes>

Changes:

[arietis27] [BEAM-13517] Unable to write nulls to columns with logical types


------------------------------------------
[...truncated 126.23 KB...]
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c84f4f21-4a31-4be7-a017-c1eca2f4edfc
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7b419bcd-acfc-446a-a6ef-0701e9f7c80c
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9e02e239-f8c2-4794-a5d2-a6645612a675

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-97346e09-29a1-4fa7-84b4-bd2fc2ab313d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3154e3eb-967a-4478-b84c-21a499a13607
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3c9f4b34-b4db-4c63-8fee-c6232be48d93
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-d7f683f8-d3a3-4b40-a3da-62c5b576e022

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c8042131-80bf-44d4-8ec6-76d8ff9529ec
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e61d300b-2f0d-4b59-92b2-741c72ef4663
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-20d2a00c-2bc5-4641-a869-a1a1666caa97
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7c0d21fd-1934-45dc-8e14-9ee6334048ce

> Task :runners:spark:3:validatesRunnerBatch
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_2_piece0 on localhost:35245 in memory (size: 8.1 KiB, free: 1458.5 MiB)
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_4_piece0 on localhost:35245 in memory (size: 9.3 KiB, free: 1458.5 MiB)
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1d26bedf-8d96-4f08-991f-67b178a3ebb7
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9d7e9334-9a91-4758-a575-98a2299b5892
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-b142d546-e46c-49a9-911b-348dd67a4e72
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f795cd35-428e-4c87-aba8-d330334a27ab
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-88941813-665c-455e-afad-8ef045886bfc

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-5b9e2914-1bf5-48f8-a883-5b79dac3f41e

> Task :runners:spark:3:validatesRunnerBatch
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_33_piece0 on localhost:45253 in memory (size: 13.9 KiB, free: 1458.3 MiB)
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_19_piece0 on localhost:45253 in memory (size: 9.7 KiB, free: 1458.3 MiB)
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0b47e905-4c04-4893-80c4-d8c1534bc6bf

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-36c345d9-d6ac-4932-9adb-ea7ba5f35847

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ba7395f9-88d9-45fb-b7e6-c596780a356b
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-73f1d394-a6c2-4f45-978c-62173b2fb5ae
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6f42f58d-ae01-423b-8eb8-f4c3ae71d1b0

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1249f072-1569-4500-a201-376495090b4d

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-5bd10f9a-3426-4c00-8d80-a8d0adb2d365
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a95bacf3-1de6-4e32-a067-caf2f57bd818
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-5922a5f5-f6e7-4990-a718-df8a48c066a4

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c253e3da-a8cf-4587-8193-118048af6b7f

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-d509aa7c-4ce6-4130-b55e-d2c45411f5fe

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 12m 33s
104 actionable tasks: 57 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/m266wmvpv73s4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9753

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9753/display/redirect?page=changes>

Changes:

[stranniknm] [BEAM-13501] playground - add log about processing and precompiled run

[noreply] [BEAM-12865] Fix triggering_frequency validation (#16315)

[noreply] [BEAM-13525] Sickbay new PardoTests for Dataflow V2 Java Streaming VR

[noreply] [BEAM-12830] Avoid flock and prepare task once. (OSX fix.) (#16333)

[noreply] Bump Python beam-master containers (#16332)


------------------------------------------
[...truncated 120.36 KB...]
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4458a800-0d4c-42fb-b209-ae72f5f3d7b7
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-74fad35c-f8be-48b3-b59b-266ed22d0eb7
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-2fe4a0b5-41ff-456d-89d8-cd8ac4e34fed
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f72010ad-a6e4-4147-ba92-eb6ccc954d30

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8d4dbefc-7975-40ae-af7a-75aad777713c
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7aa2bb4a-71f1-4366-8233-ecad9df97065
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3a5aad7f-9743-4df0-aceb-f5769607e4dd
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-841caff9-1079-4d29-a4e1-a842da847ed7

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-002e9372-d89b-4665-9fa3-a6b15ccdf9de
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-987c5fb3-9fe8-4614-ba54-72db8bfdfdaa
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f322ff4d-6d37-456d-943e-2d7a1ce1e6c5
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0b4e3738-6a39-4230-b204-f1b868eb9517

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6c6f9ae3-711b-443c-9910-224f16ba4f1f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4e312af3-c40d-4b77-85cc-eb33b98a508f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-50e6fb95-8549-46a8-b010-e6ad14e624b8
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-04d9b2f7-d298-4b01-9f38-789cac9e8059

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-631b6e37-aeac-44fc-966b-99076771ac83

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e3e179cf-f3fa-41d1-bf36-f307de31ed08

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ba5b5061-2a8b-417a-84a7-10664cc2ac8b

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-cba04193-51f3-40ac-91b8-ba649668f6b7

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4de4fde3-f121-4193-85bd-6e7fb392466a

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-56bca5bd-cf7f-4244-8d23-3206cbc1fa12
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-bf51353b-aff7-4be9-b5a0-c6814baa1bfd

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f74abe64-3d2b-48c7-a869-da2a5653f4d7
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8972a177-4069-4b4d-8435-5d2bccf7488c
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-664704ad-00b0-481d-8a65-7ecf39a02e84

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-23111815-351f-4561-9cce-84e7b697daef

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-688a653b-f55a-4216-8982-ca529d55e712

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 16m 37s
104 actionable tasks: 101 executed, 3 from cache

Publishing build scan...
https://gradle.com/s/f5utegw4ofh6y

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9752

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9752/display/redirect?page=changes>

Changes:

[noreply] [Playground][BEAM-13358]Deploy Go service via Ci/CD (#16210)

[dpcollins] [BEAM-13430] Clean up tests that override the time of the JVM

[noreply] [BEAM-12830] Install go version w/host platform. (#16330)

[noreply] [BEAM-13399] Add extra time buffer to reduce likelihood of flake in


------------------------------------------
[...truncated 115.51 KB...]
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-89e90af5-869d-460e-a882-d2dd3677197f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6a22f97b-2ed4-4b7c-a6d6-c5ddaacf789d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-d9eb30e7-8c0d-41d2-9b22-9a65d4918480

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a56e960c-6a74-4c39-8dec-284f2e718e3c
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1b7e9b83-66fc-4a72-ae90-9c41e9ab1525
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-2f52607d-2dbb-4d90-921d-1f3fc455e8b0
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-bc651e77-de2e-490c-ab90-bb1869229f7a

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4a3f2b86-5519-4cb8-b526-74a35c1c79fb
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-bf4f22f5-9ffb-4f4f-9ef1-9d030521fbf9
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6ccfaee9-a15a-4efb-834c-341f8745c398

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-78a89873-48b7-4e1a-80ea-692d0f9baa8b
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_1_piece0 on localhost:40863 in memory (size: 14.4 KiB, free: 1458.5 MiB)
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f85e21a8-c049-47f6-b8f4-d67a743420e3
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-12983ba4-eaf5-4c02-8a2c-362c8fb79c16
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-80a3d271-0573-4c79-90fd-ad27691825be

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9cf24384-b621-4728-9b8c-869ab955c3f9
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-2e66cd63-acaa-48c5-9619-b1259e72b9f0

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1d88e516-c257-443f-84c4-4f212fa7582b
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-eab0d6a4-befe-4d32-95e0-40a686e455e2

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1db797e9-dfa4-4ead-b8ac-461e83265a03

281 tests completed, 4 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-113b8451-4bf8-4221-8838-6b6e5bde52f7

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-24d68b14-2483-476f-b344-bc51b2661bf6

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c313d8d8-5d8e-4e4e-b7d2-60b65ce0e5d1
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-198e878d-cfef-43a3-a84a-de1505d8e6d8

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f0c07e67-ae7c-405d-92a5-1b7cc5a21de9
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-59d9a639-7328-4f98-adeb-4a1362a86f84
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4e90201b-1380-4a19-9480-d992912d0794

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c9d6af28-7337-45ec-8690-cbd292e33c7d

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-95385b17-8e3b-4bff-bc87-af960bc621a2

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 13m 24s
104 actionable tasks: 55 executed, 49 from cache

Publishing build scan...
https://gradle.com/s/ymy3z5x4twdtq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9751

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9751/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13494] Avoid draining Kinesis API limits & CPU when reading from

[arietis27] [BEAM-13517] Unable to write nulls to columns with logical types

[noreply] Make S3 streaming more efficient (#15931)


------------------------------------------
[...truncated 123.42 KB...]
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f5000c30-3765-4a14-bcb0-28022703fd90

> Task :runners:spark:3:validatesRunnerBatch
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9bd61b8b-fa74-4cc6-b575-574a97f4e0b7

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-43b3ee3e-a108-4a26-ac37-3c38b5934030
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-2db29edb-5d65-4061-86a2-b83b0c967c14

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ca336a5e-717c-4f5b-a7cd-fdc40bfe236f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8428458a-6def-439e-a50a-4aaeba480313
[task-result-getter-3] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 2.0 in stage 5.0 (TID 19) (localhost executor driver): TaskKilled (Stage cancelled)
[task-result-getter-3] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3e265b77-0d00-4c43-afbd-74399564b4da

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-73d331f7-fd6f-4517-8f91-dfe0a2d46070
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0c5e163d-65d4-4476-9d70-d43860e8917b
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8df541f8-4ddd-40d5-900c-8fcbfb5c0880

> Task :runners:spark:3:validatesRunnerBatch
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_2_piece0 on localhost:35223 in memory (size: 8.1 KiB, free: 1458.5 MiB)
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8752dddb-2b59-46c7-a131-d4f015550464
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-abce8fe1-75c2-4b42-8f39-28962d89c778
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1d720c65-5682-41f6-aec4-e560448fb1c6
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c687141a-8ca3-427b-bc14-56fca1dd647a

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3479c99d-9375-4043-8c52-ba3949a64736
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c491a057-0c8b-424c-8a5e-182ce80eb0df

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ec5c9f2f-326b-45ec-8b1f-1912e8120eeb
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-22f8c837-a993-41d9-a7e6-d68665e9681b

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c65509d2-a8f9-4eb4-b975-f51a6a7dc3fa

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e8f3fb82-a837-40c2-932d-4a3e3ff977d6

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a9dba90b-df16-45cf-8e1e-51d8974494a5

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c509fd24-6ae1-4976-90ff-88f834ba00b0
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-dfd544b1-dca3-45c4-a1f5-e9d3fc56ed71

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ceccee9e-9252-4205-8ffe-ba131890d409
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ba111177-94ad-4bbd-ab75-34f2dcb4fc34
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e00e617e-45c7-4f63-bb2f-a8d5925f3e9a

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-22630e92-e2fc-4271-b75f-33eae8f1c5c9

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-05c8f63e-aac7-4497-98a2-cc3714e0461d

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 18m 32s
104 actionable tasks: 59 executed, 45 from cache

Publishing build scan...
https://gradle.com/s/j4erokg4qiyqg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9750

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9750/display/redirect?page=changes>

Changes:

[Robert Bradshaw] Add microbenchmark for row coder.

[Robert Bradshaw] Compile RowCoderImpl with Cython.

[Robert Bradshaw] RowCoder optimizations.

[Robert Bradshaw] Further optimizations for common case of components in order.

[Robert Bradshaw] Optimize rows with null fields.

[Robert Bradshaw] RowCoder lint.


------------------------------------------
[...truncated 128.79 KB...]
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-5ff1fa89-40d9-4cb9-9604-a29fb11d1f05
[task-result-getter-3] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 1.0 in stage 5.0 (TID 18, localhost, executor driver): TaskKilled (Stage cancelled)
[task-result-getter-3] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3c0df9a7-f1fb-4232-90dd-d54c5b29efa1

> Task :runners:spark:3:validatesRunnerBatch
[task-result-getter-0] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 3.0 in stage 15.0 (TID 64) (localhost executor driver): TaskKilled (Stage cancelled)
[task-result-getter-0] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 15.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-2d83a1d9-e853-4677-abdc-aaa400b578de
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3e1044ab-4003-4690-b9db-1c3348e194d0
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9399f858-1dcc-4a6b-9320-1085e0a7da2c

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-32fdb742-c200-4fea-b5af-66d3b9bbafde
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-dbba2aba-f123-4966-9784-957c4b50036c
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1725a1b6-ca7b-4aef-938e-a860bb6b2d3c
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7f625227-0d1c-4ca8-978b-5659f92808ac

> Task :runners:spark:3:validatesRunnerBatch
[task-result-getter-1] INFO org.apache.spark.scheduler.TaskSetManager - Lost task 2.0 in stage 5.0 (TID 19) on localhost, executor driver: org.apache.beam.sdk.util.UserCodeException (java.lang.RuntimeException: test error in finalize) [duplicate 2]
[task-result-getter-1] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0f9fb3e4-875f-42a0-ba78-e59c8f14d72f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-84f3ff6d-2ce2-4026-8675-92339af19b25
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-bf53b67e-45ec-4cb6-b8f4-346eecba36d5
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-b62af46d-da6b-4929-a630-adc80f09cffc
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-57f9ad10-941a-4737-99b6-76fa61c78982

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9ed7e4a0-db2b-4c91-af34-963552ef316b
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-2b871b51-a3d9-4498-952b-691948aa0b5d

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4a2974dd-d7ca-48d3-aaca-3e471a820f9a
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8d09c052-1555-4ee5-8cb3-1ee2a6ed69c1

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a254101a-3416-4a6d-9964-704979e1b5da

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-9cc26070-61ab-4a82-9f9b-9c6cb7b02fac
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-81bb8bc8-aaaa-453f-9022-aa4f50157dfa

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c624acd2-828d-4eb4-bcc0-0750787ee242

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-cf462fbc-6cd6-4a92-a6cc-0bc6bed367ef

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-b87268e1-c5ec-4a3f-b129-c983e0355033
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-37114cb8-7ef1-4de4-9d63-f56b58fafb67
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4b30db36-add7-4f84-917e-7279881fcc8d

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7cbe0829-a6cb-4eb4-8046-e8b510b0f22f

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-eb4c22c2-d410-4693-ba28-89a935bed1e1

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 13m 1s
104 actionable tasks: 55 executed, 49 from cache

Publishing build scan...
https://gradle.com/s/hkak45ga4szfy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9749

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9749/display/redirect>

Changes:


------------------------------------------
[...truncated 113.10 KB...]
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-5c1cf394-eb89-45bb-b8da-7f22a80b6388
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ac0e7580-8a2d-41a6-b76c-59ef12d9c698
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a8341436-a7cb-4c6b-b98e-30cd7acc4112
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e9856646-60f1-4d07-b7d5-df5c37133be8

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a2c0af42-f919-410f-86c5-522b2cad012d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1c61c586-766c-45b5-b530-bb2aedf84620

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-063606a0-1a2a-4e37-b03d-9b9891826c6d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-b61da0d3-b065-4369-a05a-8aeeef5592ce
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0fd6b2f5-f4a5-4ed4-bd82-194ab513f1fb

> Task :runners:spark:3:validatesRunnerBatch
[task-result-getter-3] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 1.0 in stage 5.0 (TID 18) (localhost executor driver): TaskKilled (Stage cancelled)
[task-result-getter-3] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e7541586-9d7d-4fbb-8117-b06ca2194103
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f60bb6ce-2233-4c49-9ce3-76b772237c92
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-d2c5998b-3d76-4ccb-bf83-927d6380790f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c5434a49-3045-43f5-a3ab-2130e83f523a
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8ff59c26-6c40-4ece-ad17-ea274f7b70b1

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-de5afd6c-596e-469e-b17b-815ea20976db
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f954f05f-98ba-4c74-aadc-7bdba1bb3978

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-390f76f1-fb7a-43b6-a675-217a8c788c0f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f31d7606-1bb4-4469-af89-ef850d2ff54f

> Task :runners:spark:2:validatesRunnerBatch
[SessionTracker] INFO org.apache.zookeeper.server.SessionTrackerImpl - SessionTrackerImpl exited loop!
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0f6d7c52-c0c7-411b-83a6-b6558575acb4

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-add99910-bcc0-4b14-94ac-53bc1500e3f0
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-d0bc5074-426b-4465-ad4b-4b4fa41f0fb9
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8ab1d924-e5a0-448a-bd2e-5372a0bbce9e

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-00232669-505e-41e1-8e8d-706d0cb995f9

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f07683d0-2d93-487c-a6ab-1ff6a53dbe45
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-334eefb0-4601-4b3b-b818-f2ecea5d10b2
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-61ca14f4-4775-4187-8b79-77a4d3394d77

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-cd76ae8f-d265-462e-97da-56f2d5004607

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-fc946b39-40d6-4893-a5f2-4736889c774d

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 18s
104 actionable tasks: 66 executed, 38 from cache

Publishing build scan...
https://gradle.com/s/xkixw56tnewv4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9748

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9748/display/redirect?page=changes>

Changes:

[rsl4] [BEAM-13397] Bump numpy to 1.21 for M1 chip

[rsl4] re-generate requirements.txt

[Kyle Weaver] [BEAM-13498] Sickbay portable Flink testProcessElementSkew.

[noreply] Update flink cluster to use a supported dataproc version (1.2 -> 1.5)

[noreply] [BEAM-13499] Add warning about hcatalog to release notes and javadoc

[noreply] [BEAM-13430] Remove propdeps and replace with compileOnly (#16308)

[noreply] Merge pull request #16304 from [BEAM-13491] [Playground] Examples'

[noreply] Merge pull request #16283 from [BEAM-13448] [Playground] track run code

[noreply] Merge pull request #16244 from [BEAM-13463] [Playground] add retries to

[noreply] Merge pull request #16241 from [BEAM-13440] [Playground] Implement

[noreply] [BEAM-13399] Add integration test for Go SDK expansion service JAR

[noreply] [BEAM-13421] Fix bug with xs called with non-tuple key (#16258)


------------------------------------------
[...truncated 125.87 KB...]
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6290a3f9-714f-4500-8517-3fce609e8009
[task-result-getter-1] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 0.0 in stage 5.0 (TID 17, localhost, executor driver): TaskKilled (Stage cancelled)
[task-result-getter-1] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-228edf66-a3ca-463a-9a89-1a57e76cc23d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-b99f71c5-3916-4c5b-a81e-bc027cef9ac6

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ec8be1cd-9faf-442d-9bdd-26ffe3cbf81f
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-1f63a37f-c7b4-45d4-9860-332969fb0857

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-dc99464e-c584-4fc9-8189-883882080c1d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-5f306c5d-70ae-49c5-8e4a-36aca03014a3

> Task :runners:spark:3:validatesRunnerBatch
[task-result-getter-2] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-233fb1f1-2532-40c6-b378-d1629da8a0c7
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-BlockManagerMaster] INFO org.apache.spark.storage.BlockManagerInfo - Removed broadcast_9_piece0 on localhost:42213 in memory (size: 11.2 KiB, free: 1458.5 MiB)
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-efcbb2d4-c5e1-4a38-9d59-017b06731915

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8b72a46e-cdf4-43f0-a76d-bdf12604f664

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e04fc560-3c6d-44ca-8ead-86a1b411c279
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a5a8f274-d7db-4b7a-bde6-7dd1f2c4051d
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7cbb4bcd-59d9-4447-ba67-bba9c4123d19

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-ff02ddf9-9287-472f-a602-62959063f352
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3bfd705e-ad5f-4c56-8df2-bdb9071506c3

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-316374c3-7e24-4a14-99bf-29ab477d21ae
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-73af6415-2c25-4556-acf0-a34000a210c8

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-a5328aa1-89fc-41ed-b0f2-a4ca4e02a2f3

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3198c6f2-65bd-42b0-abcb-984617adc347

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4e0c4ff0-50d6-49c9-a4e9-a4063fa8f377

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f5780d39-6872-407b-b262-100eb626e339
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-fed0b175-26c7-4030-83cc-3ff772a02582

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-e7d305e4-92ae-4409-96cb-da0f1606fef0
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-39989938-ab78-4191-9b89-94a0b0636784
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8e4efe0b-c783-419f-a916-01c73b20259f

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3c89da34-cf98-46b3-8026-848f1b6769cc

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3d6758fb-1aff-4b2a-93ca-8c755703b7f3

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 21m 14s
104 actionable tasks: 62 executed, 42 from cache

Publishing build scan...
https://gradle.com/s/4hjcdluyge6m2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9747

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9747/display/redirect?page=changes>

Changes:

[stranniknm] [BEAM-13502]: fix loading example on embedded version

[noreply] [BEAM-13430] Start the process of upgrading the Gradle 7. (#16292)

[noreply] [BEAM-13430] Introduce new testRuntimeMigration configuration that


------------------------------------------
[...truncated 128.70 KB...]
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7fa3288e-4da6-4140-bee4-79e16b545235
[Executor task launch worker for task 18] INFO org.apache.spark.executor.Executor - Executor killed task 1.0 in stage 5.0 (TID 18), reason: Stage cancelled
[task-result-getter-3] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 1.0 in stage 5.0 (TID 18, localhost, executor driver): TaskKilled (Stage cancelled)
[task-result-getter-3] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-97ea5dc9-033a-4e83-8108-09975617cd68
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-4380985d-dfa3-4b06-834b-a2a191e9d90c

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-7c99ed97-a3b8-4ce9-b578-6fb3a0eaa42b
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-3f5d7403-4a44-4678-9bff-f8df32e1b2d8

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8dcd4e12-e982-4bc0-9b67-c79bd7fab4c1

> Task :runners:spark:3:validatesRunnerBatch
[Executor task launch worker for task 1.0 in stage 5.0 (TID 18)] INFO org.apache.spark.executor.Executor - Executor killed task 1.0 in stage 5.0 (TID 18), reason: Stage cancelled
[task-result-getter-0] WARN org.apache.spark.scheduler.TaskSetManager - Lost task 1.0 in stage 5.0 (TID 18) (localhost executor driver): TaskKilled (Stage cancelled)
[task-result-getter-0] INFO org.apache.spark.scheduler.TaskSchedulerImpl - Removed TaskSet 5.0, whose tasks have all completed, from pool 
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-87df015f-06cc-4835-8284-755859c899b5

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-435f8b83-0c34-4554-bd71-e5eced83444e
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6ee7b02f-dfc6-4c5b-82c2-3869c2115540

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-bff29ab7-2546-476e-9359-7013738a4569
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-cd1b4abe-de15-4cc2-81e4-e6d0761bd14c
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-abd037b6-ea65-432f-a564-f5bcd205365d

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-6be772fe-ce78-4969-9ce5-f76c8e2eadaa

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-1] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-1] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c8652e48-2eee-4539-99c8-6cc88f2a8761

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-0] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-0] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-07640746-b88b-45b1-9d12-ec27c6081b90

> Task :runners:spark:3:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-8d7be6c7-3839-47cd-b8cb-831065008a64
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from shutdown hook
[dispatcher-event-loop-2] INFO org.apache.spark.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore - MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster - BlockManagerMaster stopped
[dispatcher-event-loop-2] INFO org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-62263e61-efb8-4963-a4ab-1a87578a0264

> Task :runners:spark:2:validatesRunnerBatch
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-cad16eba-7d03-499d-9c91-662fa7a9acb2

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-bfff2cf8-1d6b-4069-b83e-d1be74f73810

> Task :runners:spark:3:validatesRunnerBatch
[SessionTracker] INFO org.apache.zookeeper.server.SessionTrackerImpl - SessionTrackerImpl exited loop!
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-d03c63be-9766-4303-9fc8-cd6b7bb8c1f6

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-f959285f-761a-450a-8dfb-b93f2f2e6c45
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-b0601ea8-901d-4280-b34b-bf1ae656af49

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-49d1df00-e960-46ae-9f01-22f0ee37a3c6
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-faf45913-a189-4d02-a4e7-5e2a0e025451
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-0321b557-e16c-4659-8821-22b701547da0

> Task :runners:spark:2:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-c2d43f61-323c-40c2-94cb-15bc0d001df2

> Task :runners:spark:3:validatesRunnerStreaming
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting directory /tmp/spark-38684780-b588-4b40-b24c-d3a8b2c3e643

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 29m 5s
104 actionable tasks: 74 executed, 30 from cache

Publishing build scan...
https://gradle.com/s/2nngmuuv3bxvq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9746

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9746/display/redirect?page=changes>

Changes:

[noreply] [BEAM-12830] Replace GoGradle plugin with Shell Scripts. (#16291)


------------------------------------------
[...truncated 24.34 KB...]
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :runners:local-java:jar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:extensions:arrow:jar
> Task :examples:java:twitter:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :runners:core-java:compileTestJava
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar

> Task :runners:core-java:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:core-java:testClasses
> Task :runners:core-java:testJar
> Task :runners:java-job-service:compileJava

> Task :sdks:java:io:hadoop-format:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:io:hadoop-format:testClasses

> Task :runners:java-job-service:compileJava
Note: <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/java-job-service/src/main/java/org/apache/beam/runners/jobsubmission/PortablePipelineJarCreator.java> uses unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:java-job-service:classes
> Task :runners:java-job-service:jar
> Task :runners:spark:2:compileJava
> Task :runners:spark:3:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:classes

> Task :runners:spark:2:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:jar
> Task :runners:spark:2:classes
> Task :runners:spark:2:jar

> Task :runners:spark:2:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:testClasses
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :runners:spark:3:validatesRunnerBatch
> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 58s
132 actionable tasks: 92 executed, 40 from cache

Publishing build scan...
https://gradle.com/s/w3ywcu4iyabps

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9745

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9745/display/redirect?page=changes>

Changes:

[benjamin.gonzalez] [BEAM-13088] Add parameter tempWriteDataset to BigQueryIO to store temp

[benjamin.gonzalez] [BEAM-13088] Make tempDataset final

[Kyle Weaver] [BEAM-13496] Upgrade Flink runner to include log4j patches.

[Kyle Weaver] [BEAM-13497] Correct class name in Flink tests.

[noreply] [BEAM-13434] Bump solr to 8.11.1 (#16296)

[noreply] Use a patched shadow 6.1.0 plugin using Log4j 2.16.0 (#16269)


------------------------------------------
[...truncated 32.36 KB...]
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE

> Task :sdks:java:io:common:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:core:compileTestJava
> Task :examples:java:twitter:compileJava
> Task :sdks:java:io:jdbc:compileJava
> Task :sdks:java:testing:test-utils:compileJava
> Task :sdks:java:io:common:testClasses
> Task :sdks:java:io:common:testJar
> Task :sdks:java:io:hadoop-format:compileJava
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :examples:java:twitter:classes
> Task :examples:java:twitter:jar
> Task :sdks:java:testing:test-utils:classes
> Task :sdks:java:testing:test-utils:jar

> Task :examples:java:twitter:compileTestJava
Note: <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/examples/java/twitter/src/test/java/org/apache/beam/examples/twitterstreamgenerator/ReadFromTwitterDoFnTest.java> uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:extensions:ml:compileJava
> Task :runners:spark:2:compileJava
> Task :runners:spark:3:compileJava
> Task :examples:java:twitter:testClasses
> Task :sdks:java:testing:test-utils:compileTestJava
> Task :sdks:java:testing:test-utils:testClasses
> Task :examples:java:twitter:testJar
> Task :sdks:java:testing:test-utils:testJar

> Task :sdks:java:io:jdbc:compileJava
Note: <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/sdks/java/io/jdbc/src/main/java/org/apache/beam/sdk/io/jdbc/JdbcIO.java> uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :sdks:java:io:hadoop-format:classes
> Task :sdks:java:io:jdbc:classes
> Task :sdks:java:extensions:ml:classes
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:jdbc:jar

> Task :sdks:java:core:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :sdks:java:core:testClasses
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :examples:java:compileJava

> Task :runners:core-java:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:core-java:testClasses
> Task :runners:core-java:testJar

> Task :examples:java:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :examples:java:classes
> Task :examples:java:jar

> Task :examples:java:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :examples:java:testClasses
> Task :examples:java:testJar

> Task :runners:spark:2:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:2:classes
> Task :runners:spark:2:jar

> Task :runners:spark:3:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:classes
> Task :runners:spark:3:jar

> Task :sdks:java:io:hadoop-format:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:io:hadoop-format:testClasses

> Task :runners:spark:2:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:testClasses
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :runners:spark:3:validatesRunnerBatch
> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

org.apache.beam.sdk.transforms.ParDoLifecycleTest > testTeardownCalledAfterExceptionInFinishBundle FAILED
    java.lang.AssertionError at ParDoLifecycleTest.java:301

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 4 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 17m 37s
132 actionable tasks: 102 executed, 30 from cache

Publishing build scan...
https://gradle.com/s/tybyzrei5qcdc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9744

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9744/display/redirect?page=changes>

Changes:

[stranniknm] [BEAM-13466]: sort categories and examples by name

[noreply] [BEAM-13399] Add check for dev versions of JARs to download code

[noreply] Merge pull request #16122 from [BEAM-13345] [Playground] add resizable

[noreply] Merge pull request #16170 from [BEAM-13411][Playground] Add getting of

[noreply] Merge pull request #16172 from [BEAM-13417] [Playground] Add java

[noreply] Merge pull request #16240 from [BEAM-13465][Playground] Change error

[noreply] Merge pull request #16234 from [BEAM-13461][Playground] [Bugfix] Error

[noreply] Merge pull request #15489 from [BEAM-12865] Allow custom batch_duration

[noreply] [BEAM-13483] Increase timeout of Java Examples Dataflow suite. (#16226)

[noreply] Pin transitive log4j dependencies to 2.17.0 in :sdks:java:io:hcatalog


------------------------------------------
[...truncated 15.99 KB...]
> Task :runners:local-java:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :examples:java:processTestResources
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :model:job-management:processResources
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:copySourceOverrides
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copySourceOverrides
> Task :sdks:java:io:common:jar
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:processResources
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processTestResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:common:testJar
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:testing:test-utils:jar
> Task :examples:java:twitter:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :examples:java:jar
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

org.apache.beam.sdk.transforms.ParDoLifecycleTest > testTeardownCalledAfterExceptionInFinishBundle FAILED
    java.lang.AssertionError at ParDoLifecycleTest.java:301

281 tests completed, 4 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 54s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/67lkzffdpnnto

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9743

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9743/display/redirect?page=changes>

Changes:

[aydar.zaynutdinov] [BEAM-13485][Playground]

[alexander.zhuravlev] [BEAM-13476] Changed Timeout Error Text

[alexander.zhuravlev] [BEAM-13474] Changed 'Playground' logo text color

[mmack] [BEAM-13443] Avoid blocking put to Kinesis records queue to shutdown

[noreply] Merge pull request #16278 from [BEAM-13479] [Playground] Change logic

[noreply] Merge pull request #16281 from [BEAM-13475] [Playground] [Bugfix] Error

[noreply] Merge pull request #16279 from [BEAM-13473] [Playground] [Bugfix] Reset

[noreply] Merge pull request #16232 from [BEAM-13285][Playground] Add steps to


------------------------------------------
[...truncated 18.19 KB...]
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :runners:local-java:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copySourceOverrides
> Task :sdks:java:io:common:jar
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:3:processResources
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:3:jar
> Task :runners:spark:2:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:3:testJar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:twitter:compileJava
> Task :examples:java:twitter:classes
> Task :examples:java:twitter:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar

> Task :examples:java:twitter:compileTestJava
Note: <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/examples/java/twitter/src/test/java/org/apache/beam/examples/twitterstreamgenerator/ReadFromTwitterDoFnTest.java> uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :examples:java:twitter:testClasses
> Task :examples:java:twitter:testJar

> Task :runners:spark:2:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :sdks:java:io:hadoop-format:compileJava
> Task :sdks:java:io:hadoop-format:classes
> Task :sdks:java:io:hadoop-format:jar

> Task :sdks:java:io:hadoop-format:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:testClasses
> Task :runners:spark:2:testJar
> Task :runners:spark:3:validatesRunnerBatch
> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 21m 35s
132 actionable tasks: 90 executed, 42 from cache

Publishing build scan...
https://gradle.com/s/xzmqwzikf76dc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9742

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9742/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13233] Replace AWS API used to list shards from DescribeStream to


------------------------------------------
[...truncated 14.96 KB...]
> Task :runners:core-java:processTestResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:direct-java:createCheckerFrameworkManifest
> Task :examples:java:processTestResources
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :runners:local-java:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :model:fn-execution:processResources
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:io:common:jar
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :model:job-management:processResources
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :runners:spark:2:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :model:pipeline:generateProto FROM-CACHE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:classes
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :model:job-management:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :examples:java:twitter:jar
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch FAILED

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 13m 21s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/ivg3vdnq7lljg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9741

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9741/display/redirect>

Changes:


------------------------------------------
[...truncated 15.19 KB...]
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :examples:java:processTestResources
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :runners:local-java:jar
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:io:hadoop-common:jar
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :examples:java:twitter:jar
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:kafka:jar
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:testJar
> Task :runners:spark:3:testJar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 21m 8s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/cjpcxb7tr6wjk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9740

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9740/display/redirect>

Changes:


------------------------------------------
[...truncated 15.16 KB...]
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :runners:direct-java:createCheckerFrameworkManifest
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :examples:java:processTestResources
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :runners:spark:2:copySourceOverrides
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:3:processResources
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:io:common:testJar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :examples:java:twitter:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 14m 21s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/cqntd6vln3cbc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9739

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9739/display/redirect>

Changes:


------------------------------------------
[...truncated 15.19 KB...]
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :runners:direct-java:createCheckerFrameworkManifest
> Task :examples:java:processTestResources
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :model:job-management:processResources
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:extractProto
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :sdks:java:io:common:jar
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :examples:java:twitter:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:2:testJar
> Task :runners:spark:3:testJar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 14m 44s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/giywkfqca2zmg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9738

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9738/display/redirect>

Changes:


------------------------------------------
[...truncated 15.20 KB...]
> Task :runners:core-java:processTestResources NO-SOURCE
> Task :sdks:java:extensions:arrow:processResources NO-SOURCE
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :examples:java:processTestResources
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:2:processResources
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:3:processResources
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :examples:java:twitter:jar
> Task :sdks:java:io:common:testJar
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:java-job-service:jar
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 29s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/fuinpsfq5bsgi

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9737

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9737/display/redirect>

Changes:


------------------------------------------
[...truncated 15.17 KB...]
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :examples:java:processTestResources
> Task :model:job-management:processResources
> Task :sdks:java:extensions:protobuf:extractProto
> Task :model:fn-execution:processResources
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :examples:java:twitter:jar
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:io:common:testJar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :sdks:java:harness:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:3:jar
> Task :runners:spark:2:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:testClasses
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 23m 11s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/4kdxcvazctagy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9736

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9736/display/redirect>

Changes:


------------------------------------------
[...truncated 15.19 KB...]
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :examples:java:processTestResources
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :model:job-management:processResources
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:copySourceOverrides
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :runners:spark:3:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:testing:test-utils:jar
> Task :runners:local-java:jar
> Task :examples:java:twitter:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:jar
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :examples:java:jar
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 12m 25s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/tdmkww53zk52i

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9735

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9735/display/redirect>

Changes:


------------------------------------------
[...truncated 15.19 KB...]
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :examples:java:processTestResources
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :model:fn-execution:processResources
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :model:job-management:processResources
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:copySourceOverrides
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :examples:java:twitter:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:common:testJar
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:jar
> Task :runners:spark:2:jar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:testClasses
> Task :examples:java:testJar
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 59s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/ekmoywmfgb5oc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9734

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9734/display/redirect>

Changes:


------------------------------------------
[...truncated 15.23 KB...]
> Task :runners:direct-java:createCheckerFrameworkManifest
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :runners:local-java:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:createCheckerFrameworkManifest
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :examples:java:processTestResources
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :model:job-management:processResources
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :model:fn-execution:processResources
> Task :runners:spark:2:copySourceOverrides
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processTestResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :runners:local-java:jar
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :examples:java:twitter:jar
> Task :sdks:java:io:hadoop-format:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:2:testJar
> Task :runners:spark:3:testJar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 14m 2s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/4gfnapo5w5mhm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9733

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9733/display/redirect>

Changes:


------------------------------------------
[...truncated 15.12 KB...]
> Task :sdks:java:io:kafka:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:extractProto
> Task :runners:java-job-service:createCheckerFrameworkManifest
> Task :runners:direct-java:createCheckerFrameworkManifest
> Task :sdks:java:io:google-cloud-platform:createCheckerFrameworkManifest
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :runners:spark:2:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :examples:java:processTestResources
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:io:hadoop-format:createCheckerFrameworkManifest
> Task :examples:java:twitter:createCheckerFrameworkManifest
> Task :sdks:java:io:jdbc:createCheckerFrameworkManifest
> Task :sdks:java:io:common:createCheckerFrameworkManifest
> Task :sdks:java:testing:test-utils:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-format:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :runners:spark:3:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:io:jdbc:processResources NO-SOURCE
> Task :examples:java:twitter:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :examples:java:twitter:processTestResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :runners:spark:2:copySourceOverrides
> Task :runners:spark:2:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:io:hadoop-format:processTestResources
> Task :runners:spark:2:createCheckerFrameworkManifest
> Task :sdks:java:io:hadoop-common:createCheckerFrameworkManifest
> Task :runners:spark:2:processResources
> Task :sdks:java:io:hadoop-common:processResources NO-SOURCE
> Task :runners:spark:2:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:copySourceOverrides
> Task :runners:spark:3:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:jar
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:jar
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:extractIncludeProto
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:arrow:jar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :runners:local-java:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:testing:test-utils:jar
> Task :examples:java:twitter:compileJava FROM-CACHE
> Task :examples:java:twitter:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :examples:java:twitter:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:testing:test-utils:testJar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-format:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :runners:core-java:testJar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :examples:java:twitter:compileTestJava FROM-CACHE
> Task :examples:java:twitter:testClasses UP-TO-DATE
> Task :examples:java:twitter:testJar
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:3:compileJava FROM-CACHE
> Task :runners:spark:3:classes
> Task :runners:spark:2:jar
> Task :runners:spark:3:jar
> Task :examples:java:compileJava FROM-CACHE
> Task :examples:java:classes UP-TO-DATE
> Task :runners:spark:3:compileTestJava FROM-CACHE
> Task :runners:spark:3:testClasses
> Task :runners:spark:2:compileTestJava FROM-CACHE
> Task :runners:spark:2:testClasses
> Task :runners:spark:3:testJar
> Task :runners:spark:2:testJar
> Task :examples:java:jar
> Task :examples:java:compileTestJava FROM-CACHE
> Task :examples:java:testClasses
> Task :examples:java:testJar
> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
> Task :sdks:java:io:hadoop-format:testClasses
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 13m 21s
132 actionable tasks: 85 executed, 47 from cache

Publishing build scan...
https://gradle.com/s/yvmbod45wszdc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Java_ValidatesRunner_Spark #9732

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/9732/display/redirect>

Changes:


------------------------------------------
[...truncated 25.05 KB...]
> Task :runners:spark:3:createCheckerFrameworkManifest
> Task :runners:spark:2:processTestResources
> Task :runners:spark:3:processResources
> Task :runners:spark:3:copyTestSourceOverrides NO-SOURCE
> Task :runners:spark:3:processTestResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto FROM-CACHE
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:pipeline:shadowJar FROM-CACHE
> Task :model:job-management:generateProto FROM-CACHE
> Task :model:fn-execution:generateProto FROM-CACHE
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:shadowJar FROM-CACHE
> Task :model:fn-execution:shadowJar FROM-CACHE
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:io:common:compileTestJava FROM-CACHE
> Task :sdks:java:io:common:testClasses UP-TO-DATE
> Task :sdks:java:extensions:arrow:compileJava FROM-CACHE
> Task :sdks:java:extensions:arrow:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:io:common:testJar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:extensions:arrow:jar
> Task :runners:local-java:jar
> Task :sdks:java:io:jdbc:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:io:jdbc:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:testing:test-utils:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:jdbc:jar
> Task :sdks:java:testing:test-utils:compileTestJava FROM-CACHE
> Task :sdks:java:testing:test-utils:testClasses UP-TO-DATE
> Task :sdks:java:testing:test-utils:testJar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:core:compileTestJava FROM-CACHE
> Task :sdks:java:core:testClasses
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:core:shadowTestJar FROM-CACHE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :runners:core-java:compileTestJava FROM-CACHE
> Task :runners:core-java:testClasses UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:testJar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar FROM-CACHE
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:direct-java:shadowJar FROM-CACHE
> Task :sdks:java:extensions:ml:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:ml:classes UP-TO-DATE
> Task :sdks:java:extensions:ml:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:spark:2:compileJava FROM-CACHE
> Task :runners:spark:2:classes
> Task :runners:spark:2:jar
> Task :sdks:java:io:google-cloud-platform:jar
> Task :examples:java:twitter:compileJava
> Task :examples:java:twitter:classes
> Task :examples:java:twitter:jar

> Task :examples:java:twitter:compileTestJava
Note: <https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/examples/java/twitter/src/test/java/org/apache/beam/examples/twitterstreamgenerator/ReadFromTwitterDoFnTest.java> uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :examples:java:twitter:testClasses
> Task :examples:java:twitter:testJar
> Task :sdks:java:io:hadoop-format:compileJava
> Task :sdks:java:io:hadoop-format:classes
> Task :sdks:java:io:hadoop-format:jar

> Task :runners:spark:2:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:compileJava
> Task :examples:java:compileJava
> Task :runners:spark:2:testClasses
> Task :runners:spark:2:testJar

> Task :examples:java:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :examples:java:classes
> Task :examples:java:jar

> Task :examples:java:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :examples:java:testClasses
> Task :examples:java:testJar

> Task :sdks:java:io:hadoop-format:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:io:hadoop-format:testClasses

> Task :runners:spark:3:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:classes
> Task :runners:spark:3:jar

> Task :runners:spark:3:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:spark:3:testClasses
> Task :runners:spark:3:testJar
> Task :runners:spark:2:validatesRunnerBatch
> Task :runners:spark:3:validatesRunnerBatch

> Task :runners:spark:2:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:3:validatesRunnerBatch

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:280

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testSingleCallOrderingWithShuffle FAILED
    java.lang.AssertionError at PerKeyOrderingTest.java:133
        Caused by: java.lang.AssertionError at MatcherAssert.java:18

org.apache.beam.sdk.transforms.PerKeyOrderingTest > testMultipleStatefulOrderingWithoutShuffle FAILED
    java.lang.IllegalStateException at PerKeyOrderingTest.java:321

> Task :runners:spark:2:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:2:validatesRunnerBatch FAILED
> Task :runners:spark:2:validatesRunnerStreaming

> Task :runners:spark:3:validatesRunnerBatch

281 tests completed, 3 failed, 1 skipped

> Task :runners:spark:3:validatesRunnerBatch FAILED
> Task :runners:spark:3:validatesRunnerStreaming

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/2/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:validatesRunnerBatch'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/ws/src/runners/spark/3/build/reports/tests/validatesRunnerBatch/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 20m 35s
132 actionable tasks: 94 executed, 38 from cache

Publishing build scan...
https://gradle.com/s/7or2wjaw4uutq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org