You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/11/02 00:27:39 UTC

Build failed in Jenkins: beam_PerformanceTests_HadoopFormat #2234

See <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/2234/display/redirect>

Changes:


------------------------------------------
[...truncated 306.94 KB...]
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Collect read time as step s2
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Get values only/Values/Map as step s3
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Values as string as step s4
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/WithKeys/AddKeys/Map as step s5
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Combine.perKey(Hashing)/GroupByKey as step s6
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues as step s7
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Values/Values/Map as step s8
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) as step s9
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/View.AsIterable/CreateDataflowView as step s10
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/CreateVoid/Read(CreateSource) as step s11
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/ProduceDefault as step s12
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Flatten.PCollections as step s13
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) as step s14
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) as step s15
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Create.Values/Read(CreateSource) as step s16
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Flatten.PCollections as step s17
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Window.Into()/Flatten.PCollections as step s18
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/WithKeys/AddKeys/Map as step s19
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/GroupByKey as step s20
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Values/Values/Map as step s21
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/ParDo(Concat) as step s22
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GetPane/Map as step s23
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/RunChecks as step s24
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/VerifyAssertions/ParDo(DefaultConclude) as step s25
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Staging pipeline description to gs://dataflow-staging-us-central1-844138762903/temp/staging/
    Nov 02, 2020 12:24:48 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading <213492 bytes, hash dee31080f1322a6c1e8a40ccfd36382df21c84bea1583bd0d5ec32b2efa9664e> to gs://dataflow-staging-us-central1-844138762903/temp/staging/pipeline-3uMQgPEyKmweikDM_TY4LfIchL6hWDvQ1ewysu-pZk4.pb
    Nov 02, 2020 12:24:49 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Dataflow SDK version: 2.26.0-SNAPSHOT
    Nov 02, 2020 12:24:50 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2020-11-01_16_24_49-14442568941178007544?project=apache-beam-testing
    Nov 02, 2020 12:24:50 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Submitted job: 2020-11-01_16_24_49-14442568941178007544
    Nov 02, 2020 12:24:50 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2020-11-01_16_24_49-14442568941178007544
    Nov 02, 2020 12:24:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    WARNING: 2020-11-02T00:24:53.414Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: hadoopformatioit0writeandreadusinghadoopformat-jenkins-110-mge1. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:09.085Z: Worker configuration: n1-standard-1 in us-central1-f.
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:09.851Z: Expanding CoGroupByKey operations into optimizable parts.
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:09.957Z: Expanding GroupByKey operations into optimizable parts.
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:09.984Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.131Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.165Z: Elided trivial flatten 
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.198Z: Unzipping flatten s17 for input s15.org.apache.beam.sdk.values.PCollection.<init>:400#d928e938ba0454ea
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.233Z: Fusing unzipped copy of PAssert$0/GroupGlobally/WithKeys/AddKeys/Map, through flatten PAssert$0/GroupGlobally/Flatten.PCollections, into producer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.277Z: Unzipping flatten s13 for input s12.org.apache.beam.sdk.values.PCollection.<init>:400#805b93b0a398202d
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.300Z: Fusing unzipped copy of PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous), through flatten Calculate hashcode/Flatten.PCollections, into producer Calculate hashcode/ProduceDefault
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.331Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Write into PAssert$0/GroupGlobally/GroupByKey/Reify
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.356Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/GroupByWindow into PAssert$0/GroupGlobally/GroupByKey/Read
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.388Z: Fusing consumer PAssert$0/GroupGlobally/Values/Values/Map into PAssert$0/GroupGlobally/GroupByKey/GroupByWindow
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.422Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(Concat) into PAssert$0/GroupGlobally/Values/Values/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.444Z: Fusing consumer PAssert$0/GetPane/Map into PAssert$0/GroupGlobally/ParDo(Concat)
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.470Z: Fusing consumer PAssert$0/RunChecks into PAssert$0/GetPane/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.501Z: Fusing consumer PAssert$0/VerifyAssertions/ParDo(DefaultConclude) into PAssert$0/RunChecks
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.525Z: Unzipping flatten s17-u31 for input s19.org.apache.beam.sdk.values.PCollection.<init>:400#f0f805bd79288b-c29
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.574Z: Fusing unzipped copy of PAssert$0/GroupGlobally/GroupByKey/Reify, through flatten PAssert$0/GroupGlobally/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.599Z: Unzipping flatten s13-u36 for input s14.org.apache.beam.sdk.values.PCollection.<init>:400#4f24fdb6be16210b-c34
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.624Z: Fusing unzipped copy of PAssert$0/GroupGlobally/ParDo(ToSingletonIterables), through flatten Calculate hashcode/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.645Z: Fusing consumer Calculate hashcode/ProduceDefault into Calculate hashcode/CreateVoid/Read(CreateSource)
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.667Z: Fusing consumer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) into Calculate hashcode/Values/Values/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.701Z: Fusing consumer Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) into Calculate hashcode/Values/Values/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.725Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) into PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.747Z: Fusing consumer Collect read time into Read using Hadoop InputFormat/Read(HadoopInputFormatBoundedSource)
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.769Z: Fusing consumer Get values only/Values/Map into Collect read time
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.805Z: Fusing consumer Values as string into Get values only/Values/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.836Z: Fusing consumer Calculate hashcode/WithKeys/AddKeys/Map into Values as string
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.859Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial into Calculate hashcode/WithKeys/AddKeys/Map
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.892Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.924Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.958Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read
    Nov 02, 2020 12:25:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:10.993Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.024Z: Fusing consumer Calculate hashcode/Values/Values/Map into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.058Z: Fusing consumer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map into PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.091Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Reify into PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.580Z: Executing operation PAssert$0/GroupGlobally/GroupByKey/Create
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.614Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.657Z: Starting 5 ****s in us-central1-f...
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.712Z: Finished operation PAssert$0/GroupGlobally/GroupByKey/Create
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.727Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.842Z: Executing operation PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:11.881Z: Executing operation Read using Hadoop InputFormat/Read(HadoopInputFormatBoundedSource)+Collect read time+Get values only/Values/Map+Values as string+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write
    Nov 02, 2020 12:25:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:12.961Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
    Nov 02, 2020 12:25:35 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:35.378Z: Autoscaling: Raised the number of ****s to 3 based on the rate of progress in the currently running stage(s).
    Nov 02, 2020 12:25:37 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:35.408Z: Resized **** pool to 3, though goal was 5.  This could be a quota issue.
    Nov 02, 2020 12:25:41 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:40.658Z: Autoscaling: Raised the number of ****s to 5 based on the rate of progress in the currently running stage(s).
    Nov 02, 2020 12:25:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:54.779Z: Workers have started successfully.
    Nov 02, 2020 12:25:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:25:54.812Z: Workers have started successfully.
    Nov 02, 2020 12:26:17 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:17.386Z: Finished operation PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 12:26:29 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:27.886Z: Finished operation Read using Hadoop InputFormat/Read(HadoopInputFormatBoundedSource)+Collect read time+Get values only/Values/Map+Values as string+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write
    Nov 02, 2020 12:26:29 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:27.965Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Close
    Nov 02, 2020 12:26:29 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:28.031Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Close
    Nov 02, 2020 12:26:29 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:28.087Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract+Calculate hashcode/Values/Values/Map+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 12:26:36 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:36.553Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract+Calculate hashcode/Values/Values/Map+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 12:26:37 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:36.715Z: Executing operation Calculate hashcode/View.AsIterable/CreateDataflowView
    Nov 02, 2020 12:26:37 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:36.765Z: Finished operation Calculate hashcode/View.AsIterable/CreateDataflowView
    Nov 02, 2020 12:26:37 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:36.897Z: Executing operation Calculate hashcode/CreateVoid/Read(CreateSource)+Calculate hashcode/ProduceDefault+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 12:26:41 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:39.759Z: Finished operation Calculate hashcode/CreateVoid/Read(CreateSource)+Calculate hashcode/ProduceDefault+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 12:26:41 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:39.880Z: Executing operation PAssert$0/GroupGlobally/GroupByKey/Close
    Nov 02, 2020 12:26:41 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:39.933Z: Finished operation PAssert$0/GroupGlobally/GroupByKey/Close
    Nov 02, 2020 12:26:41 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:39.993Z: Executing operation PAssert$0/GroupGlobally/GroupByKey/Read+PAssert$0/GroupGlobally/GroupByKey/GroupByWindow+PAssert$0/GroupGlobally/Values/Values/Map+PAssert$0/GroupGlobally/ParDo(Concat)+PAssert$0/GetPane/Map+PAssert$0/RunChecks+PAssert$0/VerifyAssertions/ParDo(DefaultConclude)
    Nov 02, 2020 12:26:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:45.593Z: Finished operation PAssert$0/GroupGlobally/GroupByKey/Read+PAssert$0/GroupGlobally/GroupByKey/GroupByWindow+PAssert$0/GroupGlobally/Values/Values/Map+PAssert$0/GroupGlobally/ParDo(Concat)+PAssert$0/GetPane/Map+PAssert$0/RunChecks+PAssert$0/VerifyAssertions/ParDo(DefaultConclude)
    Nov 02, 2020 12:26:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:45.743Z: Cleaning up.
    Nov 02, 2020 12:26:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:26:45.830Z: Stopping **** pool...
    Nov 02, 2020 12:27:30 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:27:29.781Z: Autoscaling: Resized **** pool from 5 to 0.
    Nov 02, 2020 12:27:30 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T00:27:29.826Z: Worker pool stopped.
    Nov 02, 2020 12:27:36 AM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
    INFO: Job 2020-11-01_16_24_49-14442568941178007544 finished with status DONE.

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT > writeAndReadUsingHadoopFormat STANDARD_OUT
    Load test results for test (ID): d12f6433-4d49-479a-837e-4cc9c87b44df and timestamp: 2020-11-02T00:27:36.229000000Z:
                     Metric:                    Value:
                   read_time                     9.186

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT > writeAndReadUsingHadoopFormat FAILED
    java.lang.NoSuchMethodError: org.apache.commons.lang3.StringUtils.isNoneBlank([Ljava/lang/CharSequence;)Z
        at org.apache.beam.sdk.testutils.publishing.InfluxDBPublisher.publishWithCheck(InfluxDBPublisher.java:71)
        at org.apache.beam.sdk.testutils.publishing.InfluxDBPublisher.publishWithSettings(InfluxDBPublisher.java:65)
        at org.apache.beam.sdk.testutils.metrics.IOITMetrics.publishToInflux(IOITMetrics.java:87)
        at org.apache.beam.sdk.testutils.metrics.IOITMetrics.publishToInflux(IOITMetrics.java:77)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.collectAndPublishMetrics(HadoopFormatIOIT.java:226)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.writeAndReadUsingHadoopFormat(HadoopFormatIOIT.java:210)

Gradle Test Executor 1 finished executing tests.

> Task :sdks:java:io:hadoop-format:integrationTest FAILED

1 test completed, 1 failed
Finished generating test XML results (0.024 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.031 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/reports/tests/integrationTest>
:sdks:java:io:hadoop-format:integrationTest (Thread[Execution **** for ':' Thread 9,5,main]) completed. Took 6 mins 26.468 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:hadoop-format:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/reports/tests/integrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 7m 12s
95 actionable tasks: 57 executed, 38 from cache

Publishing build scan...
https://gradle.com/s/ydgmy7jhx2u54

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PerformanceTests_HadoopFormat #2237

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/2237/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PerformanceTests_HadoopFormat #2236

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/2236/display/redirect>

Changes:


------------------------------------------
[...truncated 226.79 KB...]
:examples:java:compileTestJava (Thread[Execution **** for ':' Thread 8,5,main]) completed. Took 0.144 secs.
:examples:java:testClasses (Thread[Execution **** for ':' Thread 8,5,main]) started.

> Task :examples:java:testClasses
Skipping task ':examples:java:testClasses' as it has no actions.
:examples:java:testClasses (Thread[Execution **** for ':' Thread 8,5,main]) completed. Took 0.0 secs.
:examples:java:testJar (Thread[Execution **** for ':' Thread 8,5,main]) started.

> Task :examples:java:testJar
Caching disabled for task ':examples:java:testJar' because:
  Caching has not been enabled for the task
Task ':examples:java:testJar' is not up-to-date because:
  No history is available.
:examples:java:testJar (Thread[Execution **** for ':' Thread 8,5,main]) completed. Took 0.022 secs.

> Task :runners:direct-java:shadowJar
Could not read file path '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/direct-java/build/resources/main'.>
Custom actions are attached to task ':runners:direct-java:shadowJar'.
Caching disabled for task ':runners:direct-java:shadowJar' because:
  Caching has not been enabled for the task
Task ':runners:direct-java:shadowJar' is not up-to-date because:
  No history is available.
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/direct-java/build/resources/main',> not found
*******************
GRADLE SHADOW STATS

Total Jars: 6 (includes project)
Total Time: 0.658s [658ms]
Average Time/Jar: 0.1096666666667s [109.6666666667ms]
*******************
:runners:direct-java:shadowJar (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.919 secs.
:sdks:java:io:google-cloud-platform:compileTestJava (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :runners:google-cloud-dataflow-java:****:legacy-****:compileJava FROM-CACHE
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/src/main/java',> not found
Custom actions are attached to task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava'.
Build cache key for task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava' is 871845356925bc3579a3025bbc73f2d0
Task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava' is not up-to-date because:
  No history is available.
Loaded cache entry for task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava' with cache key 871845356925bc3579a3025bbc73f2d0
:runners:google-cloud-dataflow-java:****:legacy-****:compileJava (Thread[Execution **** for ':' Thread 9,5,main]) completed. Took 0.35 secs.
:runners:google-cloud-dataflow-java:****:legacy-****:classes (Thread[Execution **** for ':' Thread 9,5,main]) started.

> Task :runners:google-cloud-dataflow-java:****:legacy-****:classes UP-TO-DATE
Skipping task ':runners:google-cloud-dataflow-java:****:legacy-****:classes' as it has no actions.
:runners:google-cloud-dataflow-java:****:legacy-****:classes (Thread[Execution **** for ':' Thread 9,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:****:legacy-****:shadowJar (Thread[Execution **** for ':' Thread 9,5,main]) started.

> Task :sdks:java:io:google-cloud-platform:compileTestJava FROM-CACHE
Custom actions are attached to task ':sdks:java:io:google-cloud-platform:compileTestJava'.
Build cache key for task ':sdks:java:io:google-cloud-platform:compileTestJava' is dd99e219afe04ca113fa9d38d238fd81
Task ':sdks:java:io:google-cloud-platform:compileTestJava' is not up-to-date because:
  No history is available.
Loaded cache entry for task ':sdks:java:io:google-cloud-platform:compileTestJava' with cache key dd99e219afe04ca113fa9d38d238fd81
:sdks:java:io:google-cloud-platform:compileTestJava (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.46 secs.
:sdks:java:io:google-cloud-platform:testClasses (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :sdks:java:io:google-cloud-platform:testClasses
Skipping task ':sdks:java:io:google-cloud-platform:testClasses' as it has no actions.
:sdks:java:io:google-cloud-platform:testClasses (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.0 secs.
:sdks:java:io:google-cloud-platform:testJar (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :sdks:java:io:google-cloud-platform:testJar
Caching disabled for task ':sdks:java:io:google-cloud-platform:testJar' because:
  Caching has not been enabled for the task
Task ':sdks:java:io:google-cloud-platform:testJar' is not up-to-date because:
  No history is available.
:sdks:java:io:google-cloud-platform:testJar (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.32 secs.
:runners:google-cloud-dataflow-java:compileTestJava (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :runners:google-cloud-dataflow-java:compileTestJava FROM-CACHE
Custom actions are attached to task ':runners:google-cloud-dataflow-java:compileTestJava'.
Build cache key for task ':runners:google-cloud-dataflow-java:compileTestJava' is 7d1e393fd53afd27e82a6538cef72458
Task ':runners:google-cloud-dataflow-java:compileTestJava' is not up-to-date because:
  No history is available.
Loaded cache entry for task ':runners:google-cloud-dataflow-java:compileTestJava' with cache key 7d1e393fd53afd27e82a6538cef72458
:runners:google-cloud-dataflow-java:compileTestJava (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.215 secs.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :runners:google-cloud-dataflow-java:testClasses UP-TO-DATE
Skipping task ':runners:google-cloud-dataflow-java:testClasses' as it has no actions.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :runners:google-cloud-dataflow-java:testJar
Could not read file path '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/build/resources/test'.>
Caching disabled for task ':runners:google-cloud-dataflow-java:testJar' because:
  Caching has not been enabled for the task
Task ':runners:google-cloud-dataflow-java:testJar' is not up-to-date because:
  No history is available.
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/build/resources/test',> not found
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.161 secs.

> Task :runners:google-cloud-dataflow-java:****:legacy-****:shadowJar
Could not read file path '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/resources/main'.>
Could not read file path '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/original_sources_to_package'.>
Custom actions are attached to task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar'.
Caching disabled for task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar' because:
  Caching has not been enabled for the task
Task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar' is not up-to-date because:
  No history is available.
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/resources/main',> not found
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/original_sources_to_package',> not found
*******************
GRADLE SHADOW STATS

Total Jars: 16 (includes project)
Total Time: 3.397s [3397ms]
Average Time/Jar: 0.2123125s [212.3125ms]
*******************
:runners:google-cloud-dataflow-java:****:legacy-****:shadowJar (Thread[Execution **** for ':' Thread 9,5,main]) completed. Took 4.429 secs.
:sdks:java:io:hadoop-format:compileTestJava (Thread[Execution **** for ':' Thread 9,5,main]) started.

> Task :sdks:java:io:hadoop-format:compileTestJava FROM-CACHE
Custom actions are attached to task ':sdks:java:io:hadoop-format:compileTestJava'.
Build cache key for task ':sdks:java:io:hadoop-format:compileTestJava' is ac779f18bee677d1907413826efbbc4c
Task ':sdks:java:io:hadoop-format:compileTestJava' is not up-to-date because:
  No history is available.
Loaded cache entry for task ':sdks:java:io:hadoop-format:compileTestJava' with cache key ac779f18bee677d1907413826efbbc4c
:sdks:java:io:hadoop-format:compileTestJava (Thread[Execution **** for ':' Thread 9,5,main]) completed. Took 0.544 secs.
:sdks:java:io:hadoop-format:testClasses (Thread[Execution **** for ':' Thread 9,5,main]) started.

> Task :sdks:java:io:hadoop-format:testClasses
Skipping task ':sdks:java:io:hadoop-format:testClasses' as it has no actions.
:sdks:java:io:hadoop-format:testClasses (Thread[Execution **** for ':' Thread 9,5,main]) completed. Took 0.0 secs.
:sdks:java:io:hadoop-format:integrationTest (Thread[Execution **** for ':' Thread 11,5,main]) started.

> Task :sdks:java:io:hadoop-format:integrationTest
Custom actions are attached to task ':sdks:java:io:hadoop-format:integrationTest'.
Build cache key for task ':sdks:java:io:hadoop-format:integrationTest' is 4657d942fbae9efe46f8b606fe4a23c4
Task ':sdks:java:io:hadoop-format:integrationTest' is not up-to-date because:
  Task.upToDateWhen is false.
Starting process 'Gradle Test Executor 1'. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format> Command: /usr/lib/jvm/java-8-openjdk-amd64/bin/java -DbeamTestPipelineOptions=["--tempRoot=gs://temp-storage-for-perf-tests","--project=apache-beam-testing","--runner=DataflowRunner","--numberOfRecords=600000","--bigQueryDataset=beam_performance","--bigQueryTable=hadoopformatioit_results","--influxMeasurement=hadoopformatioit_results","--influxDatabase=beam_test_metrics","--influxHost=http://10.128.0.96:8086","--postgresUsername=postgres","--postgresPassword=uuinkks","--postgresDatabaseName=postgres","--postgresServerName=34.123.89.149","--postgresSsl=false","--postgresPort=5432","--numWorkers=5","--autoscalingAlgorithm=NONE","--****HarnessContainerImage=","--dataflowWorkerJar=<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.26.0-SNAPSHOT.jar","--region=us-central1"]> -Djava.security.manager=****.org.gradle.process.internal.****.child.BootstrapSecurityManager -Dorg.gradle.native=false -Xmx2g -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -ea -cp /home/jenkins/.gradle/caches/6.6.1/****Main/gradle-****.jar ****.org.gradle.process.internal.****.GradleWorkerMain 'Gradle Test Executor 1'
Successfully started process 'Gradle Test Executor 1'

Gradle Test Executor 1 started executing tests.

> Task :sdks:java:io:hadoop-format:integrationTest

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT STANDARD_ERROR
    SLF4J: Class path contains multiple SLF4J bindings.
    SLF4J: Found binding in [jar:<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.26.0-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-jdk14/1.7.30/d35953dd2fe54ebe39fdf18cfd82fe6eb35b25ed/slf4j-jdk14-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/ch.qos.logback/logback-classic/1.1.3/d90276fff414f06cb375f2057f6778cd63c6082f/logback-classic-1.1.3.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-log4j12/1.7.30/c21f55139d8141d2231214fb1feaf50a1edca95e/slf4j-log4j12-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
    SLF4J: Actual binding is of type [org.slf4j.impl.JDK14LoggerFactory]
    Nov 02, 2020 12:20:25 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Attempt #1 of 3 failed: The connection attempt failed..
    Nov 02, 2020 12:20:25 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Retrying in 2000 ms.
    Nov 02, 2020 12:20:37 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Attempt #2 of 3 failed: The connection attempt failed..
    Nov 02, 2020 12:20:37 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Retrying in 4000 ms.
    Nov 02, 2020 12:20:51 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Attempt #3 of 3 failed: The connection attempt failed..
    Nov 02, 2020 12:21:01 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Attempt #1 of 3 failed: The connection attempt failed..
    Nov 02, 2020 12:21:01 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Retrying in 2000 ms.
    Nov 02, 2020 12:21:03 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Attempt #2 of 3 failed: ERROR: table "beamtest_hadoopformatioit_2020_11_02_12_20_15_202" does not exist.
    Nov 02, 2020 12:21:03 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Retrying in 4000 ms.
    Nov 02, 2020 12:21:07 PM org.apache.beam.sdk.io.common.IOITHelper executeWithRetry
    WARNING: Attempt #3 of 3 failed: ERROR: table "beamtest_hadoopformatioit_2020_11_02_12_20_15_202" does not exist.

Gradle Test Executor 1 finished executing tests.

> Task :sdks:java:io:hadoop-format:integrationTest FAILED

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT > classMethod FAILED
    org.postgresql.util.PSQLException: The connection attempt failed.
        at org.postgresql.core.v3.ConnectionFactoryImpl.openConnectionImpl(ConnectionFactoryImpl.java:315)
        at org.postgresql.core.ConnectionFactory.openConnection(ConnectionFactory.java:51)
        at org.postgresql.jdbc.PgConnection.<init>(PgConnection.java:225)
        at org.postgresql.Driver.makeConnection(Driver.java:465)
        at org.postgresql.Driver.connect(Driver.java:264)
        at java.sql.DriverManager.getConnection(DriverManager.java:664)
        at java.sql.DriverManager.getConnection(DriverManager.java:247)
        at org.postgresql.ds.common.BaseDataSource.getConnection(BaseDataSource.java:103)
        at org.postgresql.ds.common.BaseDataSource.getConnection(BaseDataSource.java:87)
        at org.apache.beam.sdk.io.common.DatabaseTestHelper.createTable(DatabaseTestHelper.java:45)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.createTable(HadoopFormatIOIT.java:132)
        at org.apache.beam.sdk.io.common.IOITHelper.executeWithRetry(IOITHelper.java:87)
        at org.apache.beam.sdk.io.common.IOITHelper.executeWithRetry(IOITHelper.java:67)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.setUp(HadoopFormatIOIT.java:127)

        Caused by:
        java.net.SocketTimeoutException: connect timed out
            at java.net.PlainSocketImpl.socketConnect(Native Method)
            at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
            at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
            at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
            at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
            at java.net.Socket.connect(Socket.java:607)
            at org.postgresql.core.PGStream.createSocket(PGStream.java:231)
            at org.postgresql.core.PGStream.<init>(PGStream.java:95)
            at org.postgresql.core.v3.ConnectionFactoryImpl.tryConnect(ConnectionFactoryImpl.java:98)
            at org.postgresql.core.v3.ConnectionFactoryImpl.openConnectionImpl(ConnectionFactoryImpl.java:213)
            ... 13 more

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT > classMethod FAILED
    org.postgresql.util.PSQLException: ERROR: table "beamtest_hadoopformatioit_2020_11_02_12_20_15_202" does not exist
        at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2553)
        at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2285)
        at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:323)
        at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:473)
        at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:393)
        at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:322)
        at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:308)
        at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:284)
        at org.postgresql.jdbc.PgStatement.executeUpdate(PgStatement.java:258)
        at org.apache.beam.sdk.io.common.DatabaseTestHelper.deleteTable(DatabaseTestHelper.java:65)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.deleteTable(HadoopFormatIOIT.java:172)
        at org.apache.beam.sdk.io.common.IOITHelper.executeWithRetry(IOITHelper.java:87)
        at org.apache.beam.sdk.io.common.IOITHelper.executeWithRetry(IOITHelper.java:67)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.tearDown(HadoopFormatIOIT.java:168)

2 tests completed, 2 failed
Finished generating test XML results (0.015 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.022 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/reports/tests/integrationTest>
:sdks:java:io:hadoop-format:integrationTest (Thread[Execution **** for ':' Thread 11,5,main]) completed. Took 57.502 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:hadoop-format:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/reports/tests/integrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 47s
95 actionable tasks: 57 executed, 38 from cache

Publishing build scan...
https://gradle.com/s/btf7mqgnqpqxy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PerformanceTests_HadoopFormat #2235

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/2235/display/redirect>

Changes:


------------------------------------------
[...truncated 306.04 KB...]
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Collect read time as step s2
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Get values only/Values/Map as step s3
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Values as string as step s4
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/WithKeys/AddKeys/Map as step s5
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Combine.perKey(Hashing)/GroupByKey as step s6
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues as step s7
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Values/Values/Map as step s8
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) as step s9
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/View.AsIterable/CreateDataflowView as step s10
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/CreateVoid/Read(CreateSource) as step s11
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/ProduceDefault as step s12
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Calculate hashcode/Flatten.PCollections as step s13
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) as step s14
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) as step s15
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Create.Values/Read(CreateSource) as step s16
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Flatten.PCollections as step s17
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Window.Into()/Flatten.PCollections as step s18
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/WithKeys/AddKeys/Map as step s19
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/GroupByKey as step s20
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Values/Values/Map as step s21
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/ParDo(Concat) as step s22
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GetPane/Map as step s23
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/RunChecks as step s24
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/VerifyAssertions/ParDo(DefaultConclude) as step s25
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Staging pipeline description to gs://dataflow-staging-us-central1-844138762903/temp/staging/
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading <213474 bytes, hash 7965cb4fa08ecaa287d8c0c3c6c875423c9845e3cbe5752379bbf538e9352bd6> to gs://dataflow-staging-us-central1-844138762903/temp/staging/pipeline-eWXLT6COyqKH2MDDxsh1QjyYRePL5XUjebv1OOk1K9Y.pb
    Nov 02, 2020 6:24:22 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Dataflow SDK version: 2.26.0-SNAPSHOT
    Nov 02, 2020 6:24:24 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2020-11-01_22_24_22-2064084361431052722?project=apache-beam-testing
    Nov 02, 2020 6:24:24 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Submitted job: 2020-11-01_22_24_22-2064084361431052722
    Nov 02, 2020 6:24:24 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2020-11-01_22_24_22-2064084361431052722
    Nov 02, 2020 6:24:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    WARNING: 2020-11-02T06:24:27.272Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: hadoopformatioit0writeandreadusinghadoopformat-jenkins-110-no32. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
    Nov 02, 2020 6:24:37 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:37.764Z: Worker configuration: n1-standard-1 in us-central1-f.
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.455Z: Expanding CoGroupByKey operations into optimizable parts.
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.582Z: Expanding GroupByKey operations into optimizable parts.
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.608Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.747Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.779Z: Elided trivial flatten 
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.814Z: Unzipping flatten s17 for input s15.org.apache.beam.sdk.values.PCollection.<init>:400#d928e938ba0454ea
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.846Z: Fusing unzipped copy of PAssert$0/GroupGlobally/WithKeys/AddKeys/Map, through flatten PAssert$0/GroupGlobally/Flatten.PCollections, into producer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.880Z: Unzipping flatten s13 for input s12.org.apache.beam.sdk.values.PCollection.<init>:400#805b93b0a398202d
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.904Z: Fusing unzipped copy of PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous), through flatten Calculate hashcode/Flatten.PCollections, into producer Calculate hashcode/ProduceDefault
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.940Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Write into PAssert$0/GroupGlobally/GroupByKey/Reify
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:38.985Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/GroupByWindow into PAssert$0/GroupGlobally/GroupByKey/Read
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.020Z: Fusing consumer PAssert$0/GroupGlobally/Values/Values/Map into PAssert$0/GroupGlobally/GroupByKey/GroupByWindow
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.054Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(Concat) into PAssert$0/GroupGlobally/Values/Values/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.124Z: Fusing consumer PAssert$0/GetPane/Map into PAssert$0/GroupGlobally/ParDo(Concat)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.154Z: Fusing consumer PAssert$0/RunChecks into PAssert$0/GetPane/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.180Z: Fusing consumer PAssert$0/VerifyAssertions/ParDo(DefaultConclude) into PAssert$0/RunChecks
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.204Z: Unzipping flatten s17-u31 for input s19.org.apache.beam.sdk.values.PCollection.<init>:400#f0f805bd79288b-c29
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.237Z: Fusing unzipped copy of PAssert$0/GroupGlobally/GroupByKey/Reify, through flatten PAssert$0/GroupGlobally/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.270Z: Unzipping flatten s13-u36 for input s14.org.apache.beam.sdk.values.PCollection.<init>:400#4f24fdb6be16210b-c34
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.294Z: Fusing unzipped copy of PAssert$0/GroupGlobally/ParDo(ToSingletonIterables), through flatten Calculate hashcode/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.328Z: Fusing consumer Calculate hashcode/ProduceDefault into Calculate hashcode/CreateVoid/Read(CreateSource)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.350Z: Fusing consumer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) into Calculate hashcode/Values/Values/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.375Z: Fusing consumer Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) into Calculate hashcode/Values/Values/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.432Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) into PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.464Z: Fusing consumer Collect read time into Read using Hadoop InputFormat/Read(HadoopInputFormatBoundedSource)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.508Z: Fusing consumer Get values only/Values/Map into Collect read time
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.541Z: Fusing consumer Values as string into Get values only/Values/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.565Z: Fusing consumer Calculate hashcode/WithKeys/AddKeys/Map into Values as string
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.600Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial into Calculate hashcode/WithKeys/AddKeys/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.634Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.670Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.696Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.723Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.746Z: Fusing consumer Calculate hashcode/Values/Values/Map into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.774Z: Fusing consumer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map into PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:39.804Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Reify into PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
    Nov 02, 2020 6:24:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.184Z: Executing operation PAssert$0/GroupGlobally/GroupByKey/Create
    Nov 02, 2020 6:24:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.265Z: Starting 5 ****s in us-central1-f...
    Nov 02, 2020 6:24:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.322Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
    Nov 02, 2020 6:24:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.327Z: Finished operation PAssert$0/GroupGlobally/GroupByKey/Create
    Nov 02, 2020 6:24:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.382Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
    Nov 02, 2020 6:24:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.541Z: Executing operation PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 6:24:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:40.582Z: Executing operation Read using Hadoop InputFormat/Read(HadoopInputFormatBoundedSource)+Collect read time+Get values only/Values/Map+Values as string+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write
    Nov 02, 2020 6:24:54 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:24:54.036Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
    Nov 02, 2020 6:26:01 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:00.212Z: Workers have started successfully.
    Nov 02, 2020 6:26:04 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:04.010Z: Autoscaling: Raised the number of ****s to 2 based on the rate of progress in the currently running stage(s).
    Nov 02, 2020 6:26:04 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:04.079Z: Resized **** pool to 2, though goal was 5.  This could be a quota issue.
    Nov 02, 2020 6:26:13 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:11.480Z: Workers have started successfully.
    Nov 02, 2020 6:26:20 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:19.863Z: Autoscaling: Raised the number of ****s to 5 based on the rate of progress in the currently running stage(s).
    Nov 02, 2020 6:26:29 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:27.859Z: Finished operation PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 6:26:38 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:36.314Z: Finished operation Read using Hadoop InputFormat/Read(HadoopInputFormatBoundedSource)+Collect read time+Get values only/Values/Map+Values as string+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write
    Nov 02, 2020 6:26:38 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:36.382Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Close
    Nov 02, 2020 6:26:38 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:36.438Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Close
    Nov 02, 2020 6:26:38 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:36.548Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract+Calculate hashcode/Values/Values/Map+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 6:26:47 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:46.897Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract+Calculate hashcode/Values/Values/Map+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 6:26:47 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:47.061Z: Executing operation Calculate hashcode/View.AsIterable/CreateDataflowView
    Nov 02, 2020 6:26:47 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:47.117Z: Finished operation Calculate hashcode/View.AsIterable/CreateDataflowView
    Nov 02, 2020 6:26:47 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:47.262Z: Executing operation Calculate hashcode/CreateVoid/Read(CreateSource)+Calculate hashcode/ProduceDefault+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 6:26:50 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:50.069Z: Finished operation Calculate hashcode/CreateVoid/Read(CreateSource)+Calculate hashcode/ProduceDefault+PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)+PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$0/GroupGlobally/WithKeys/AddKeys/Map+PAssert$0/GroupGlobally/GroupByKey/Reify+PAssert$0/GroupGlobally/GroupByKey/Write
    Nov 02, 2020 6:26:50 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:50.141Z: Executing operation PAssert$0/GroupGlobally/GroupByKey/Close
    Nov 02, 2020 6:26:50 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:50.196Z: Finished operation PAssert$0/GroupGlobally/GroupByKey/Close
    Nov 02, 2020 6:26:50 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:50.268Z: Executing operation PAssert$0/GroupGlobally/GroupByKey/Read+PAssert$0/GroupGlobally/GroupByKey/GroupByWindow+PAssert$0/GroupGlobally/Values/Values/Map+PAssert$0/GroupGlobally/ParDo(Concat)+PAssert$0/GetPane/Map+PAssert$0/RunChecks+PAssert$0/VerifyAssertions/ParDo(DefaultConclude)
    Nov 02, 2020 6:26:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:54.037Z: Finished operation PAssert$0/GroupGlobally/GroupByKey/Read+PAssert$0/GroupGlobally/GroupByKey/GroupByWindow+PAssert$0/GroupGlobally/Values/Values/Map+PAssert$0/GroupGlobally/ParDo(Concat)+PAssert$0/GetPane/Map+PAssert$0/RunChecks+PAssert$0/VerifyAssertions/ParDo(DefaultConclude)
    Nov 02, 2020 6:26:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:54.181Z: Cleaning up.
    Nov 02, 2020 6:26:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:26:54.255Z: Stopping **** pool...
    Nov 02, 2020 6:28:45 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:28:44.658Z: Autoscaling: Resized **** pool from 5 to 0.
    Nov 02, 2020 6:28:45 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2020-11-02T06:28:44.708Z: Worker pool stopped.
    Nov 02, 2020 6:28:50 AM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
    INFO: Job 2020-11-01_22_24_22-2064084361431052722 finished with status DONE.

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT > writeAndReadUsingHadoopFormat STANDARD_OUT
    Load test results for test (ID): 91375563-8b34-4783-a96f-56952c234458 and timestamp: 2020-11-02T06:28:50.513000000Z:
                     Metric:                    Value:
                   read_time                    14.258

org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT > writeAndReadUsingHadoopFormat FAILED
    java.lang.NoSuchMethodError: org.apache.commons.lang3.StringUtils.isNoneBlank([Ljava/lang/CharSequence;)Z
        at org.apache.beam.sdk.testutils.publishing.InfluxDBPublisher.publishWithCheck(InfluxDBPublisher.java:71)
        at org.apache.beam.sdk.testutils.publishing.InfluxDBPublisher.publishWithSettings(InfluxDBPublisher.java:65)
        at org.apache.beam.sdk.testutils.metrics.IOITMetrics.publishToInflux(IOITMetrics.java:87)
        at org.apache.beam.sdk.testutils.metrics.IOITMetrics.publishToInflux(IOITMetrics.java:77)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.collectAndPublishMetrics(HadoopFormatIOIT.java:226)
        at org.apache.beam.sdk.io.hadoop.format.HadoopFormatIOIT.writeAndReadUsingHadoopFormat(HadoopFormatIOIT.java:210)

Gradle Test Executor 1 finished executing tests.

> Task :sdks:java:io:hadoop-format:integrationTest FAILED

1 test completed, 1 failed
Finished generating test XML results (0.022 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.03 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/reports/tests/integrationTest>
:sdks:java:io:hadoop-format:integrationTest (Thread[Execution **** for ':' Thread 2,5,main]) completed. Took 8 mins 12.152 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:hadoop-format:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_HadoopFormat/ws/src/sdks/java/io/hadoop-format/build/reports/tests/integrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 8m 59s
95 actionable tasks: 57 executed, 38 from cache

Publishing build scan...
https://gradle.com/s/fgtzbi5whfvms

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org