You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/12/10 16:51:20 UTC

Build failed in Jenkins: beam_PostCommit_Java_PortabilityApi_GradleBuild #389

See <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/389/display/redirect?page=changes>

Changes:

[iemejia] [BEAM-6079] Add ability for CassandraIO to delete data

[iemejia] [BEAM-6079] Fix access level and clean up generics issues

------------------------------------------
[...truncated 1.35 MB...]
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map as step s63
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization) as step s64
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/View.CreatePCollectionView as step s65
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/rewindowIntoGlobal/Window.Assign as step s66
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WriteBundlesToFiles as step s67
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/GroupByDestination as step s68
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WriteGroupedRecords as step s69
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/FlattenFiles as step s70
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/View.AsIterable/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization) as step s71
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/View.AsIterable/View.CreatePCollectionView as step s72
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Impulse as step s73
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/ParDo(SplitBoundedSource) as step s74
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair with random key as step s75
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign as step s76
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey as step s77
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable as step s78
    Dec 10, 2018 4:51:01 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map as step s79
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/ParDo(ReadFromBoundedSource) as step s80
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/ParDo(Anonymous) as step s81
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WritePartitionUntriggered as step s82
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsReshuffle/Window.Into()/Window.Assign as step s83
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsReshuffle/GroupByKey as step s84
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsReshuffle/ExpandIterable as step s85
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/ParMultiDo(WriteTables) as step s86
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/WithKeys/AddKeys/Map as step s87
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/Window.Into()/Window.Assign as step s88
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/GroupByKey as step s89
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/Values/Values/Map as step s90
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/ParDo(GarbageCollectTemporaryFiles) as step s91
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/View.AsIterable/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization) as step s92
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/View.AsIterable/View.CreatePCollectionView as step s93
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Impulse as step s94
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/ParDo(SplitBoundedSource) as step s95
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair with random key as step s96
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign as step s97
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey as step s98
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable as step s99
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map as step s100
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/ParDo(ReadFromBoundedSource) as step s101
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/ParDo(Anonymous) as step s102
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WriteRenameUntriggered as step s103
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionsReshuffle/Window.Into()/Window.Assign as step s104
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionsReshuffle/GroupByKey as step s105
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionsReshuffle/ExpandIterable as step s106
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/ParMultiDo(WriteTables) as step s107
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/WithKeys/AddKeys/Map as step s108
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/Window.Into()/Window.Assign as step s109
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/GroupByKey as step s110
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/Values/Values/Map as step s111
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/ParDo(GarbageCollectTemporaryFiles) as step s112
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Impulse as step s113
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/ParDo(SplitBoundedSource) as step s114
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair with random key as step s115
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign as step s116
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey as step s117
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable as step s118
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map as step s119
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/ParDo(ReadFromBoundedSource) as step s120
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding DropInputs as step s121
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Staging pipeline description to gs://temp-storage-for-end-to-end-tests/testpipeline-jenkins-1210165051-2a0b53ef/output/results/staging/
    Dec 10, 2018 4:51:02 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading <348872 bytes, hash iqaSmJl_9-Hv_2DbC4CosA> to gs://temp-storage-for-end-to-end-tests/testpipeline-jenkins-1210165051-2a0b53ef/output/results/staging/pipeline-iqaSmJl_9-Hv_2DbC4CosA.pb

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom STANDARD_OUT
    Dataflow SDK version: 2.10.0-SNAPSHOT

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom STANDARD_ERROR
    Dec 10, 2018 4:51:03 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-12-10_08_51_02-1065202652686330085?project=apache-beam-testing

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom STANDARD_OUT
    Submitted job: 2018-12-10_08_51_02-1065202652686330085

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom STANDARD_ERROR
    Dec 10, 2018 4:51:03 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2018-12-10_08_51_02-1065202652686330085
    Dec 10, 2018 4:51:03 PM org.apache.beam.runners.dataflow.TestDataflowRunner run
    INFO: Running Dataflow job 2018-12-10_08_51_02-1065202652686330085 with 0 expected assertions.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-12-10T16:51:02.546Z: Autoscaling is enabled for job 2018-12-10_08_51_02-1065202652686330085. The number of workers will be between 1 and 1000.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-12-10T16:51:02.595Z: Autoscaling was automatically enabled for job 2018-12-10_08_51_02-1065202652686330085.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-12-10T16:51:05.222Z: Checking permissions granted to controller Service Account.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-12-10T16:51:09.379Z: Worker configuration: n1-standard-1 in us-central1-b.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    SEVERE: 2018-12-10T16:51:09.941Z: Workflow failed. Causes: Project apache-beam-testing has insufficient quota(s) to execute this workflow with 1 instances in region us-central1. Quota summary (required/available): 1/6987 instances, 1/0 CPUs, 250/69951 disk GB, 0/4046 SSD disk GB, 1/288 instance groups, 1/288 managed instance groups, 1/263 instance templates, 1/241 in-use IP addresses.

    Please see https://cloud.google.com/compute/docs/resource-quotas about requesting more quota.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-12-10T16:51:10.026Z: Cleaning up.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-12-10T16:51:10.084Z: Worker pool stopped.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.TestDataflowRunner$ErrorMonitorMessagesHandler process
    INFO: Dataflow job 2018-12-10_08_51_02-1065202652686330085 threw exception. Failure message was: Workflow failed. Causes: Project apache-beam-testing has insufficient quota(s) to execute this workflow with 1 instances in region us-central1. Quota summary (required/available): 1/6987 instances, 1/0 CPUs, 250/69951 disk GB, 0/4046 SSD disk GB, 1/288 instance groups, 1/288 managed instance groups, 1/263 instance templates, 1/241 in-use IP addresses.

    Please see https://cloud.google.com/compute/docs/resource-quotas about requesting more quota.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-12-10_08_51_02-1065202652686330085 failed with status FAILED.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    WARNING: Metrics not present for Dataflow job 2018-12-10_08_51_02-1065202652686330085.
    Dec 10, 2018 4:51:15 PM org.apache.beam.runners.dataflow.TestDataflowRunner run
    WARNING: Dataflow job 2018-12-10_08_51_02-1065202652686330085 did not output a success or failure metric.
    Dec 10, 2018 4:51:15 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT cleanBqEnvironment
    INFO: Start to clean up tables and datasets.
    Dec 10, 2018 4:51:15 PM org.apache.beam.sdk.io.gcp.testing.BigqueryClient deleteDataset
    INFO: Successfully deleted dataset: bq_query_to_table_1544460651649_155

Gradle Test Executor 4 finished executing tests.

> Task :beam-runners-google-cloud-dataflow-java:googleCloudPlatformFnApiWorkerIntegrationTest FAILED

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom FAILED
    java.lang.RuntimeException: Workflow failed. Causes: Project apache-beam-testing has insufficient quota(s) to execute this workflow with 1 instances in region us-central1. Quota summary (required/available): 1/6987 instances, 1/0 CPUs, 250/69951 disk GB, 0/4046 SSD disk GB, 1/288 instance groups, 1/288 managed instance groups, 1/263 instance templates, 1/241 in-use IP addresses.

    Please see https://cloud.google.com/compute/docs/resource-quotas about requesting more quota.
        at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:134)
        at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:90)
        at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:55)
        at org.apache.beam.sdk.Pipeline.run(Pipeline.java:313)
        at org.apache.beam.sdk.Pipeline.run(Pipeline.java:299)
        at org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT.runBigQueryToTablePipeline(BigQueryToTableIT.java:111)
        at org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT.testStandardQueryWithoutCustom(BigQueryToTableIT.java:295)

7 tests completed, 5 failed
Finished generating test XML results (0.028 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformFnApiWorkerIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.02 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformFnApiWorkerIntegrationTest>
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformFnApiWorkerIntegrationTest (Thread[Task worker for ':' Thread 3,5,main]) completed. Took 10 mins 50.92 secs.
:beam-runners-google-cloud-dataflow-java:cleanUpDockerImages (Thread[Task worker for ':' Thread 3,5,main]) started.

> Task :beam-runners-google-cloud-dataflow-java:cleanUpDockerImages
Caching disabled for task ':beam-runners-google-cloud-dataflow-java:cleanUpDockerImages': Caching has not been enabled for the task
Task ':beam-runners-google-cloud-dataflow-java:cleanUpDockerImages' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
Starting process 'command 'docker''. Working directory: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/ws/src/runners/google-cloud-dataflow-java> Command: docker rmi us.gcr.io/apache-beam-testing/java-postcommit-it/java:20181210163145
Successfully started process 'command 'docker''
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20181210163145
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d1dbd4e7b17be1b2e58c1ee8b77c0ad5d9166dcafcb43ddb98b7b6a9c3e6c622
Starting process 'command 'gcloud''. Working directory: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/ws/src/runners/google-cloud-dataflow-java> Command: gcloud --quiet container images delete --force-delete-tags us.gcr.io/apache-beam-testing/java-postcommit-it/java:20181210163145
Successfully started process 'command 'gcloud''
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d1dbd4e7b17be1b2e58c1ee8b77c0ad5d9166dcafcb43ddb98b7b6a9c3e6c622
  Associated tags:
 - 20181210163145
Tags:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java:20181210163145
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20181210163145].
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d1dbd4e7b17be1b2e58c1ee8b77c0ad5d9166dcafcb43ddb98b7b6a9c3e6c622].
:beam-runners-google-cloud-dataflow-java:cleanUpDockerImages (Thread[Task worker for ':' Thread 3,5,main]) completed. Took 2.587 secs.

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-runners-google-cloud-dataflow-java:examplesJavaFnApiWorkerIntegrationTest'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/examplesJavaFnApiWorkerIntegrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-runners-google-cloud-dataflow-java:googleCloudPlatformFnApiWorkerIntegrationTest'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformFnApiWorkerIntegrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/4.10.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 19m 46s
90 actionable tasks: 84 executed, 5 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/ymq4pujxwecso

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Java_PortabilityApi_GradleBuild #390

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi_GradleBuild/390/display/redirect>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org