You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/08/03 06:47:49 UTC

Build failed in Jenkins: beam_PostCommit_Java_GradleBuild #1162

See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/1162/display/redirect>

------------------------------------------
[...truncated 20.60 MB...]
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.158Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.448Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.491Z: Elided trivial flatten 
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.533Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/Wait/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Create seed/Read(CreateSource)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.569Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Read information schema into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/Wait/Map
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.610Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.654Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow) into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.692Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Read information schema
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.736Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.773Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.817Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey) into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.863Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.908Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.952Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:39.998Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.042Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/ParDo(CollectWindows) into SpannerIO.Write/To mutation group
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.076Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Extract keys into SpannerIO.Write/Write mutations to Cloud Spanner/Serialize mutations
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.121Z: Fusing consumer ParDo(GenerateMutations) into GenerateSequence/Read(BoundedCountingSource)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.156Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.197Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Batch mutations together into SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/GroupByWindow
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.237Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike) into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.272Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.317Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow) into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Extract
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.362Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Reify
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.403Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Extract into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.439Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Partial
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.484Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Partial into SpannerIO.Write/Write mutations to Cloud Spanner/Extract keys
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.524Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Reify
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.569Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.601Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/View.AsList/ParDo(ToIsmRecordForGlobalWindow) into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Flatten.Iterables/FlattenIterables/FlatMap
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.636Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.679Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Flatten.Iterables/FlattenIterables/FlatMap into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Values/Values/Map
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.712Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Values/Values/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Extract
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.754Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForKey) into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.790Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Extract into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.833Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/WithKeys/AddKeys/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/ParDo(CollectWindows)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.877Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.911Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Write mutations to Spanner into SpannerIO.Write/Write mutations to Cloud Spanner/Batch mutations together
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.945Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Partial
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:40.990Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Partial into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/WithKeys/AddKeys/Map
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.028Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Partition input
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.070Z: Fusing consumer SpannerIO.Write/To mutation group into ParDo(GenerateMutations)
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.115Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/GroupByWindow into SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.155Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Reify
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.193Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForSize) into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Read
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.657Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.702Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.747Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.758Z: Starting 1 workers in us-central1-b...
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.777Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.817Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.862Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.910Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:41.940Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Create
    Aug 03, 2018 6:43:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:42.388Z: Executing operation GenerateSequence/Read(BoundedCountingSource)+ParDo(GenerateMutations)+SpannerIO.Write/To mutation group+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/ParDo(CollectWindows)+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/WithKeys/AddKeys/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Partial+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Write
    Aug 03, 2018 6:43:53 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:43:51.306Z: Autoscaling: Raised the number of workers to 0 based on the rate of progress in the currently running step(s).
    Aug 03, 2018 6:44:04 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:01.945Z: Autoscaling: Raised the number of workers to 1 based on the rate of progress in the currently running step(s).
    Aug 03, 2018 6:44:04 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:02.055Z: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job.
    Aug 03, 2018 6:44:19 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:18.507Z: Workers have started successfully.
    Aug 03, 2018 6:44:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:39.436Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Close
    Aug 03, 2018 6:44:40 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:39.579Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Extract+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Values/Values/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Flatten.Iterables/FlattenIterables/FlatMap+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/View.AsList/ParDo(ToIsmRecordForGlobalWindow)
    Aug 03, 2018 6:44:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:48.358Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/View.AsList/CreateDataflowView
    Aug 03, 2018 6:44:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:48.562Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Create seed/Read(CreateSource)+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/Wait/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Read information schema+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write
    Aug 03, 2018 6:44:56 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:56.490Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Close
    Aug 03, 2018 6:44:56 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:44:56.568Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
    Aug 03, 2018 6:45:01 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:01.108Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
    Aug 03, 2018 6:45:01 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:01.221Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow)
    Aug 03, 2018 6:45:10 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:09.726Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/CreateDataflowView
    Aug 03, 2018 6:45:10 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:09.936Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Serialize mutations+SpannerIO.Write/Write mutations to Cloud Spanner/Extract keys+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Partial+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Write
    Aug 03, 2018 6:45:22 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:21.702Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Close
    Aug 03, 2018 6:45:22 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:21.792Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Extract+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow)+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
    Aug 03, 2018 6:45:26 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:26.208Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
    Aug 03, 2018 6:45:26 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:26.298Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike)+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Write+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Write
    Aug 03, 2018 6:45:35 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:35.161Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Close
    Aug 03, 2018 6:45:35 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:35.198Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Close
    Aug 03, 2018 6:45:35 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:35.248Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForSize)
    Aug 03, 2018 6:45:35 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:35.299Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForKey)
    Aug 03, 2018 6:45:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:47.302Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/Flatten.PCollections
    Aug 03, 2018 6:45:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:47.563Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/CreateDataflowView
    Aug 03, 2018 6:45:48 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:47.788Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Partition input+SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Write
    Aug 03, 2018 6:45:59 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:58.897Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Close
    Aug 03, 2018 6:45:59 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:45:58.995Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/GroupByWindow+SpannerIO.Write/Write mutations to Cloud Spanner/Batch mutations together+SpannerIO.Write/Write mutations to Cloud Spanner/Write mutations to Spanner
    Aug 03, 2018 6:46:08 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:46:07.414Z: Cleaning up.
    Aug 03, 2018 6:46:08 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:46:07.716Z: Stopping worker pool...
    Aug 03, 2018 6:47:36 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:47:33.791Z: Autoscaling: Resized worker pool from 1 to 0.
    Aug 03, 2018 6:47:36 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-03T06:47:33.886Z: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job.
    Aug 03, 2018 6:47:43 AM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-08-02_23_43_29-12638636226223298224 finished with status DONE.
    Aug 03, 2018 6:47:43 AM org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    INFO: Success result for Dataflow job 2018-08-02_23_43_29-12638636226223298224. Found 0 success, 0 failures out of 0 expected assertions.
    Aug 03, 2018 6:47:43 AM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-08-02_23_43_29-12638636226223298224 finished with status DONE.

Gradle Test Executor 143 finished executing tests.

> Task :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest
Finished generating test XML results (0.021 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.022 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest>
Packing task ':beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest'
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest (Thread[Task worker for ':' Thread 6,5,main]) completed. Took 28 mins 11.286 secs.
:beam-runners-google-cloud-dataflow-java:buildDependents (Thread[Task worker for ':' Thread 8,5,main]) started.

> Task :beam-runners-google-cloud-dataflow-java:buildDependents
Caching disabled for task ':beam-runners-google-cloud-dataflow-java:buildDependents': Caching has not been enabled for the task
Task ':beam-runners-google-cloud-dataflow-java:buildDependents' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
:beam-runners-google-cloud-dataflow-java:buildDependents (Thread[Task worker for ':' Thread 8,5,main]) completed. Took 0.002 secs.
:beam-sdks-java-io-google-cloud-platform:buildDependents (Thread[Task worker for ':' Thread 8,5,main]) started.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Task worker for ':',5,main]) started.

> Task :beam-runners-google-cloud-dataflow-java:postCommit
Skipping task ':beam-runners-google-cloud-dataflow-java:postCommit' as it has no actions.

> Task :beam-sdks-java-io-google-cloud-platform:buildDependents
Caching disabled for task ':beam-sdks-java-io-google-cloud-platform:buildDependents': Caching has not been enabled for the task
Task ':beam-sdks-java-io-google-cloud-platform:buildDependents' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Task worker for ':',5,main]) completed. Took 0.0 secs.
:beam-sdks-java-io-google-cloud-platform:buildDependents (Thread[Task worker for ':' Thread 8,5,main]) completed. Took 0.0 secs.
:beam-sdks-java-extensions-protobuf:buildDependents (Thread[Task worker for ':' Thread 8,5,main]) started.

> Task :beam-sdks-java-extensions-protobuf:buildDependents
Caching disabled for task ':beam-sdks-java-extensions-protobuf:buildDependents': Caching has not been enabled for the task
Task ':beam-sdks-java-extensions-protobuf:buildDependents' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
:beam-sdks-java-extensions-protobuf:buildDependents (Thread[Task worker for ':' Thread 8,5,main]) completed. Took 0.0 secs.

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-io-elasticsearch-tests-2:compileTestJava'.
> Compilation failed with exit code 1; see the compiler error output for details.

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-io-mqtt:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/io/mqtt/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 47m 38s
748 actionable tasks: 735 executed, 13 from cache

Publishing build scan...
https://gradle.com/s/p2ex7gdl3z7xs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

Jenkins build is back to normal : beam_PostCommit_Java_GradleBuild #1164

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/1164/display/redirect?page=changes>


Build failed in Jenkins: beam_PostCommit_Java_GradleBuild #1163

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/1163/display/redirect>

------------------------------------------
[...truncated 8.90 MB...]
    Aug 03, 2018 12:12:55 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testSQLRead_2018_08_03_12_12_55_664_1751598958186675481

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testSelectsPayloadContent STANDARD_ERROR
    Aug 03, 2018 12:12:55 PM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testSelectsPayloadContent-2018-08-03-12-12-47-249-events-7515464311332386215_beam_8782202243310816060 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testSelectsPayloadContent-2018-08-03-12-12-47-249-events-7515464311332386215. Note this subscription WILL NOT be deleted when the pipeline terminates

org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryReadWriteIT > testSQLRead STANDARD_ERROR
    Aug 03, 2018 12:12:55 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testSQLRead_2018_08_03_12_12_55_821_2820406466237710712
    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    INSERT INTO `beam`.`TEST`
    VALUES ROW(9223372036854775807, 127, 32767, 2147483647, 1.0, 1.0, TRUE, TIMESTAMP '2018-05-28 20:17:40.123', 'varchar', 'char', ARRAY['123', '456'])
    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      LogicalProject(c_bigint=[9223372036854775807], c_tinyint=[127], c_smallint=[32767], c_integer=[2147483647], c_float=[1.0], c_double=[1.0], c_boolean=[true], c_timestamp=[2018-05-28 20:17:40.123], c_varchar=['varchar'], c_char=['char'], c_arr=[ARRAY('123', '456')])
        LogicalValues(tuples=[[{ 0 }]])

    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      BeamCalcRel(expr#0=[{inputs}], expr#1=[9223372036854775807], expr#2=[127], expr#3=[32767], expr#4=[2147483647], expr#5=[1.0], expr#6=[true], expr#7=[2018-05-28 20:17:40.123], expr#8=['varchar'], expr#9=['char'], expr#10=['123'], expr#11=['456'], expr#12=[ARRAY($t10, $t11)], c_bigint=[$t1], c_tinyint=[$t2], c_smallint=[$t3], c_integer=[$t4], c_float=[$t5], c_double=[$t5], c_boolean=[$t6], c_timestamp=[$t7], c_varchar=[$t8], c_char=[$t9], c_arr=[$t12])
        BeamValuesRel(tuples=[[{ 0 }]])

    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    INSERT INTO `beam`.`TEST`
    VALUES ROW(9223372036854775807, 127, 32767, 2147483647, 1.0, 1.0, TRUE, TIMESTAMP '2018-05-28 20:17:40.123', 'varchar', 'char', ARRAY['123', '456'])
    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      LogicalProject(c_bigint=[9223372036854775807], c_tinyint=[127], c_smallint=[32767], c_integer=[2147483647], c_float=[1.0], c_double=[1.0], c_boolean=[true], c_timestamp=[2018-05-28 20:17:40.123], c_varchar=['varchar'], c_char=['char'], c_arr=[ARRAY('123', '456')])
        LogicalValues(tuples=[[{ 0 }]])

    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      BeamCalcRel(expr#0=[{inputs}], expr#1=[9223372036854775807], expr#2=[127], expr#3=[32767], expr#4=[2147483647], expr#5=[1.0], expr#6=[true], expr#7=[2018-05-28 20:17:40.123], expr#8=['varchar'], expr#9=['char'], expr#10=['123'], expr#11=['456'], expr#12=[ARRAY($t10, $t11)], c_bigint=[$t1], c_tinyint=[$t2], c_smallint=[$t3], c_integer=[$t4], c_float=[$t5], c_double=[$t5], c_boolean=[$t6], c_timestamp=[$t7], c_varchar=[$t8], c_char=[$t9], c_arr=[$t12])
        BeamValuesRel(tuples=[[{ 0 }]])

    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.io.gcp.bigquery.BatchLoads$4 getTempFilePrefix
    INFO: Writing BigQuery temporary files to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87/ before loading them.
    Aug 03, 2018 12:12:56 PM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87/1013cbcc-18a2-46c3-a175-9a4ceefbcdc9.
    Aug 03, 2018 12:12:57 PM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Loading 1 files into {datasetId=integ_test, projectId=apache-beam-testing, tableId=BigQueryReadWriteIT_testSQLRead_2018_08_03_12_12_55_664_1751598958186675481} using job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0, location=US, projectId=apache-beam-testing}, attempt 1
    Aug 03, 2018 12:12:57 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl startJob
    INFO: Started BigQuery job: {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0, location=US, projectId=apache-beam-testing}.
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0
    Aug 03, 2018 12:12:57 PM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0, location=US, projectId=apache-beam-testing} started
    Aug 03, 2018 12:12:57 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0
    Aug 03, 2018 12:12:59 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 03, 2018 12:12:59 PM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0803121256ad9b01ef_ed64394375fb4e618b32a32aae489e87_8936ffe318c63e8eda3234d3a7be7cbc_00001_00000-0, location=US, projectId=apache-beam-testing} succeeded. Statistics: {"creationTime":"1533298377374","endTime":"1533298378298","load":{"badRecords":"0","inputFileBytes":"243","inputFiles":"1","outputBytes":"82","outputRows":"1"},"startTime":"1533298377549"}
    Aug 03, 2018 12:12:59 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    SELECT `TEST`.`c_bigint`, `TEST`.`c_tinyint`, `TEST`.`c_smallint`, `TEST`.`c_integer`, `TEST`.`c_float`, `TEST`.`c_double`, `TEST`.`c_boolean`, `TEST`.`c_timestamp`, `TEST`.`c_varchar`, `TEST`.`c_char`, `TEST`.`c_arr`
    FROM `beam`.`TEST` AS `TEST`
    Aug 03, 2018 12:12:59 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    LogicalProject(c_bigint=[$0], c_tinyint=[$1], c_smallint=[$2], c_integer=[$3], c_float=[$4], c_double=[$5], c_boolean=[$6], c_timestamp=[$7], c_varchar=[$8], c_char=[$9], c_arr=[$10])
      BeamIOSourceRel(table=[[beam, TEST]])

    Aug 03, 2018 12:12:59 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamCalcRel(expr#0..10=[{inputs}], proj#0..10=[{exprs}])
      BeamIOSourceRel(table=[[beam, TEST]])

    Aug 03, 2018 12:13:00 PM org.apache.beam.sdk.io.gcp.bigquery.BigQuerySourceBase executeExtract
    INFO: Starting BigQuery extract job: beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract
    Aug 03, 2018 12:13:00 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl startJob
    INFO: Started BigQuery job: {jobId=beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract, location=US, projectId=apache-beam-testing}.
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract
    Aug 03, 2018 12:13:00 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract
    Aug 03, 2018 12:13:01 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 03, 2018 12:13:01 PM org.apache.beam.sdk.io.gcp.bigquery.BigQuerySourceBase executeExtract
    INFO: BigQuery extract job completed: beam_job_ff93f9674bb64b089b811ceb82fc1302_bigqueryreadwriteit0testsqlreadjenkins080312130084b9f594-extract
    Aug 03, 2018 12:13:01 PM org.apache.beam.sdk.io.gcp.bigquery.BigQuerySourceBase split
    INFO: Extract job produced 1 files
    Aug 03, 2018 12:13:01 PM org.apache.beam.sdk.io.FileBasedSource createReader
    INFO: Matched 1 files for pattern gs://temp-storage-for-end-to-end-tests/BigQueryExtractTemp/ff93f9674bb64b089b811ceb82fc1302/000000000000.avro
    Aug 03, 2018 12:13:02 PM org.apache.beam.sdk.io.FileBasedSource getEstimatedSizeBytes
    INFO: Filepattern gs://temp-storage-for-end-to-end-tests/BigQueryExtractTemp/ff93f9674bb64b089b811ceb82fc1302/000000000000.avro matched 1 files with total size 738

org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryReadWriteIT > testInsertSelect STANDARD_ERROR
    Aug 03, 2018 12:13:02 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testInsertSelect_2018_08_03_12_13_02_907_8511757694076647395
    Aug 03, 2018 12:13:03 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testInsertSelect_2018_08_03_12_13_03_121_5096128434114429741
    Aug 03, 2018 12:13:03 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    INSERT INTO `beam`.`ORDERS_BQ`
    (SELECT `ORDERS_IN_MEMORY`.`id` AS `id`, `ORDERS_IN_MEMORY`.`name` AS `name`, `ORDERS_IN_MEMORY`.`arr` AS `arr`
    FROM `beam`.`ORDERS_IN_MEMORY` AS `ORDERS_IN_MEMORY`)
    Aug 03, 2018 12:13:03 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    BeamIOSinkRel(table=[[beam, ORDERS_BQ]], operation=[INSERT], flattened=[true])
      LogicalProject(id=[$0], name=[$1], arr=[$2])
        BeamIOSourceRel(table=[[beam, ORDERS_IN_MEMORY]])

    Aug 03, 2018 12:13:03 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamIOSinkRel(table=[[beam, ORDERS_BQ]], operation=[INSERT], flattened=[true])
      BeamCalcRel(expr#0..2=[{inputs}], proj#0..2=[{exprs}])
        BeamIOSourceRel(table=[[beam, ORDERS_IN_MEMORY]])

    Aug 03, 2018 12:13:04 PM org.apache.beam.sdk.io.gcp.bigquery.BatchLoads$4 getTempFilePrefix
    INFO: Writing BigQuery temporary files to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1/ before loading them.
    Aug 03, 2018 12:13:04 PM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1/8c9e6c77-fd38-424e-85c9-8131c65384bc.
    Aug 03, 2018 12:13:04 PM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1/4de4f2a3-7a8c-4776-b5a2-96b5557f7ba6.
    Aug 03, 2018 12:13:04 PM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1/25bb01c0-a3b9-4552-a46b-8418274b1f50.
    Aug 03, 2018 12:13:04 PM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Loading 3 files into {datasetId=integ_test, projectId=apache-beam-testing, tableId=BigQueryReadWriteIT_testInsertSelect_2018_08_03_12_13_03_121_5096128434114429741} using job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0, location=US, projectId=apache-beam-testing}, attempt 1
    Aug 03, 2018 12:13:05 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl startJob
    INFO: Started BigQuery job: {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0, location=US, projectId=apache-beam-testing}.
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0
    Aug 03, 2018 12:13:05 PM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0, location=US, projectId=apache-beam-testing} started
    Aug 03, 2018 12:13:05 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0
    Aug 03, 2018 12:13:06 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 03, 2018 12:13:06 PM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins080312130454f3c542_77bdf9d1f59c4afc8a23ec74a818b1a1_96e8fb98c51b6cdb1da9a2fd0774b2a4_00001_00000-0, location=US, projectId=apache-beam-testing} succeeded. Statistics: {"creationTime":"1533298385006","endTime":"1533298386493","load":{"badRecords":"0","inputFileBytes":"126","inputFiles":"3","outputBytes":"69","outputRows":"3"},"startTime":"1533298385338"}

Gradle Test Executor 100 finished executing tests.
Gradle Test Executor 101 finished executing tests.

> Task :beam-sdks-java-extensions-sql:integrationTest

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testUsesDlq STANDARD_ERROR
    Aug 03, 2018 12:13:19 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    SELECT `message`.`payload`.`id`, `message`.`payload`.`name`
    FROM `beam`.`message` AS `message`
    Aug 03, 2018 12:13:19 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    LogicalProject(id=[$2], name=[$3])
      LogicalProject(event_timestamp=[$0], attributes=[$1], id=[$2.id], name=[$2.name])
        BeamIOSourceRel(table=[[beam, message]])

    Aug 03, 2018 12:13:19 PM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamCalcRel(expr#0..2=[{inputs}], expr#3=[$t2.id], expr#4=[$t2.name], id=[$t3], name=[$t4])
      BeamIOSourceRel(table=[[beam, message]])

    Aug 03, 2018 12:13:24 PM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testUsesDlq-2018-08-03-12-13-18-647-events--8195838382457819063_beam_-4208529694066292862 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testUsesDlq-2018-08-03-12-13-18-647-events--8195838382457819063. Note this subscription WILL NOT be deleted when the pipeline terminates
    Aug 03, 2018 12:13:27 PM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testUsesDlq-2018-08-03-12-13-18-461-events--1728356866840818933_beam_-2943474913320426189 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testUsesDlq-2018-08-03-12-13-18-461-events--1728356866840818933. Note this subscription WILL NOT be deleted when the pipeline terminates
    Aug 03, 2018 12:13:42 PM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/start-subscription-2404237263819267391 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}
    Aug 03, 2018 12:14:04 PM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription-6477398772255271349 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}
    Aug 03, 2018 12:14:19 PM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription-6477398772255271349 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}
    Aug 03, 2018 12:14:35 PM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription-6477398772255271349 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}
    Aug 03, 2018 12:14:50 PM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription-6477398772255271349 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testUsesDlq FAILED
    java.lang.AssertionError: Did not receive signal on projects/apache-beam-testing/subscriptions/result-subscription-6477398772255271349 in 60s
        at org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal.pollForResultForDuration(TestPubsubSignal.java:269)
        at org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal.waitForSuccess(TestPubsubSignal.java:237)
        at org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT.testUsesDlq(PubsubJsonIT.java:206)

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testSQLLimit STANDARD_ERROR
    Aug 03, 2018 12:14:56 PM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testSQLLimit-2018-08-03-12-14-52-481-events-6907496877142992184_beam_-5805428676725456273 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testSQLLimit-2018-08-03-12-14-52-481-events-6907496877142992184. Note this subscription WILL NOT be deleted when the pipeline terminates

Gradle Test Executor 102 finished executing tests.

> Task :beam-sdks-java-extensions-sql:integrationTest FAILED

7 tests completed, 1 failed
Finished generating test XML results (0.002 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.01 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/reports/tests/integrationTest>
:beam-sdks-java-extensions-sql:integrationTest (Thread[Task worker for ':' Thread 10,5,main]) completed. Took 3 mins 20.251 secs.

FAILURE: Build completed with 4 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-extensions-google-cloud-platform-core:compileTestJava'.
> Compilation failed with exit code 1; see the compiler error output for details.

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-runners-core-java:compileTestJava'.
> Compilation failed with exit code 1; see the compiler error output for details.

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-io-mqtt:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/io/mqtt/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-extensions-sql:integrationTest'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/reports/tests/integrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 52s
687 actionable tasks: 674 executed, 13 from cache

Publishing build scan...
https://gradle.com/s/5aawourt6rb3o

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure