You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/08/13 12:00:14 UTC

Build failed in Jenkins: beam_PostCommit_Java_GradleBuild #1240

See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/1240/display/redirect?page=changes>

Changes:

[lukasz.gajowy] [BEAM-3359] Remove setting Flink master to [auto] in TestFlinkRunner

[lukasz.gajowy] [BEAM-3370] Allow to run IO Performance tests using Flink runner

[lukasz.gajowy] [BEAM-3370] Fix problem with directories not being staged from classpath

------------------------------------------
[...truncated 12.67 MB...]
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqltypesjenkins0813115719737b4350_bea3509a54e34b0887a33e6bd157bb4c_6523838a715c9d550c2d33ef870e0b66_00001_00000-0
    Aug 13, 2018 11:57:20 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testsqltypesjenkins0813115719737b4350_bea3509a54e34b0887a33e6bd157bb4c_6523838a715c9d550c2d33ef870e0b66_00001_00000-0, location=US, projectId=apache-beam-testing} started
    Aug 13, 2018 11:57:20 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_load_bigqueryreadwriteit0testsqltypesjenkins0813115719737b4350_bea3509a54e34b0887a33e6bd157bb4c_6523838a715c9d550c2d33ef870e0b66_00001_00000-0, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqltypesjenkins0813115719737b4350_bea3509a54e34b0887a33e6bd157bb4c_6523838a715c9d550c2d33ef870e0b66_00001_00000-0

org.apache.beam.sdk.extensions.sql.PubsubToBigqueryIT > testSimpleInsert STANDARD_ERROR
    Aug 13, 2018 11:57:21 AM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubToBigqueryIT-testSimpleInsert-2018-08-13-11-57-14-030-events--6694147617290134922_beam_1826379710737790282 to topic projects/apache-beam-testing/topics/integ-test-PubsubToBigqueryIT-testSimpleInsert-2018-08-13-11-57-14-030-events--6694147617290134922. Note this subscription WILL NOT be deleted when the pipeline terminates

org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryReadWriteIT > testSQLTypes STANDARD_ERROR
    Aug 13, 2018 11:57:22 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_load_bigqueryreadwriteit0testsqltypesjenkins0813115719737b4350_bea3509a54e34b0887a33e6bd157bb4c_6523838a715c9d550c2d33ef870e0b66_00001_00000-0, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 13, 2018 11:57:22 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testsqltypesjenkins0813115719737b4350_bea3509a54e34b0887a33e6bd157bb4c_6523838a715c9d550c2d33ef870e0b66_00001_00000-0, location=US, projectId=apache-beam-testing} succeeded. Statistics: {"creationTime":"1534161440208","endTime":"1534161441519","load":{"badRecords":"0","inputFileBytes":"243","inputFiles":"1","outputBytes":"82","outputRows":"1"},"startTime":"1534161440473"}

org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryReadWriteIT > testSQLRead STANDARD_ERROR
    Aug 13, 2018 11:57:22 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testSQLRead_2018_08_13_11_57_22_896_5596107784928585728
    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testSQLRead_2018_08_13_11_57_23_089_7699743411458827911
    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    INSERT INTO `beam`.`TEST`
    VALUES ROW(9223372036854775807, 127, 32767, 2147483647, 1.0, 1.0, TRUE, TIMESTAMP '2018-05-28 20:17:40.123', 'varchar', 'char', ARRAY['123', '456'])
    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      LogicalProject(c_bigint=[9223372036854775807], c_tinyint=[127], c_smallint=[32767], c_integer=[2147483647], c_float=[1.0], c_double=[1.0], c_boolean=[true], c_timestamp=[2018-05-28 20:17:40.123], c_varchar=['varchar'], c_char=['char'], c_arr=[ARRAY('123', '456')])
        LogicalValues(tuples=[[{ 0 }]])

    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      BeamCalcRel(expr#0=[{inputs}], expr#1=[9223372036854775807], expr#2=[127], expr#3=[32767], expr#4=[2147483647], expr#5=[1.0], expr#6=[true], expr#7=[2018-05-28 20:17:40.123], expr#8=['varchar'], expr#9=['char'], expr#10=['123'], expr#11=['456'], expr#12=[ARRAY($t10, $t11)], c_bigint=[$t1], c_tinyint=[$t2], c_smallint=[$t3], c_integer=[$t4], c_float=[$t5], c_double=[$t5], c_boolean=[$t6], c_timestamp=[$t7], c_varchar=[$t8], c_char=[$t9], c_arr=[$t12])
        BeamValuesRel(tuples=[[{ 0 }]])

    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    INSERT INTO `beam`.`TEST`
    VALUES ROW(9223372036854775807, 127, 32767, 2147483647, 1.0, 1.0, TRUE, TIMESTAMP '2018-05-28 20:17:40.123', 'varchar', 'char', ARRAY['123', '456'])
    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      LogicalProject(c_bigint=[9223372036854775807], c_tinyint=[127], c_smallint=[32767], c_integer=[2147483647], c_float=[1.0], c_double=[1.0], c_boolean=[true], c_timestamp=[2018-05-28 20:17:40.123], c_varchar=['varchar'], c_char=['char'], c_arr=[ARRAY('123', '456')])
        LogicalValues(tuples=[[{ 0 }]])

    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamIOSinkRel(table=[[beam, TEST]], operation=[INSERT], flattened=[true])
      BeamCalcRel(expr#0=[{inputs}], expr#1=[9223372036854775807], expr#2=[127], expr#3=[32767], expr#4=[2147483647], expr#5=[1.0], expr#6=[true], expr#7=[2018-05-28 20:17:40.123], expr#8=['varchar'], expr#9=['char'], expr#10=['123'], expr#11=['456'], expr#12=[ARRAY($t10, $t11)], c_bigint=[$t1], c_tinyint=[$t2], c_smallint=[$t3], c_integer=[$t4], c_float=[$t5], c_double=[$t5], c_boolean=[$t6], c_timestamp=[$t7], c_varchar=[$t8], c_char=[$t9], c_arr=[$t12])
        BeamValuesRel(tuples=[[{ 0 }]])


org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testSelectsPayloadContent STANDARD_ERROR
    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testSelectsPayloadContent-2018-08-13-11-57-15-262-events-223743193486745281_beam_670756351085949259 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testSelectsPayloadContent-2018-08-13-11-57-15-262-events-223743193486745281. Note this subscription WILL NOT be deleted when the pipeline terminates

org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryReadWriteIT > testSQLRead STANDARD_ERROR
    Aug 13, 2018 11:57:23 AM org.apache.beam.sdk.io.gcp.bigquery.BatchLoads$4 getTempFilePrefix
    INFO: Writing BigQuery temporary files to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48/ before loading them.
    Aug 13, 2018 11:57:24 AM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48/4f84a7ad-beb8-4275-a879-59832545d827.
    Aug 13, 2018 11:57:24 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Loading 1 files into {datasetId=integ_test, projectId=apache-beam-testing, tableId=BigQueryReadWriteIT_testSQLRead_2018_08_13_11_57_22_896_5596107784928585728} using job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, location=US, projectId=apache-beam-testing}, attempt 1
    Aug 13, 2018 11:57:24 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl startJob
    INFO: Started BigQuery job: {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, location=US, projectId=apache-beam-testing}.
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0
    Aug 13, 2018 11:57:24 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, location=US, projectId=apache-beam-testing} started
    Aug 13, 2018 11:57:24 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0
    Aug 13, 2018 11:57:25 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0
    Aug 13, 2018 11:57:27 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 13, 2018 11:57:27 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testsqlreadjenkins0813115723eede9d0f_13feb25de5d540b29495590578c5fc48_7c9a627de52f289899a04240bdf7d2f9_00001_00000-0, location=US, projectId=apache-beam-testing} succeeded. Statistics: {"completionRatio":1.0,"creationTime":"1534161444609","endTime":"1534161445903","load":{"badRecords":"0","inputFileBytes":"243","inputFiles":"1","outputBytes":"82","outputRows":"1"},"startTime":"1534161444856"}
    Aug 13, 2018 11:57:27 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    SELECT `TEST`.`c_bigint`, `TEST`.`c_tinyint`, `TEST`.`c_smallint`, `TEST`.`c_integer`, `TEST`.`c_float`, `TEST`.`c_double`, `TEST`.`c_boolean`, `TEST`.`c_timestamp`, `TEST`.`c_varchar`, `TEST`.`c_char`, `TEST`.`c_arr`
    FROM `beam`.`TEST` AS `TEST`
    Aug 13, 2018 11:57:27 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    LogicalProject(c_bigint=[$0], c_tinyint=[$1], c_smallint=[$2], c_integer=[$3], c_float=[$4], c_double=[$5], c_boolean=[$6], c_timestamp=[$7], c_varchar=[$8], c_char=[$9], c_arr=[$10])
      BeamIOSourceRel(table=[[beam, TEST]])

    Aug 13, 2018 11:57:28 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamCalcRel(expr#0..10=[{inputs}], proj#0..10=[{exprs}])
      BeamIOSourceRel(table=[[beam, TEST]])

    Aug 13, 2018 11:57:28 AM org.apache.beam.sdk.io.gcp.bigquery.BigQuerySourceBase executeExtract
    INFO: Starting BigQuery extract job: beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract
    Aug 13, 2018 11:57:28 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl startJob
    INFO: Started BigQuery job: {jobId=beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract, location=US, projectId=apache-beam-testing}.
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract
    Aug 13, 2018 11:57:28 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract
    Aug 13, 2018 11:57:30 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 13, 2018 11:57:30 AM org.apache.beam.sdk.io.gcp.bigquery.BigQuerySourceBase executeExtract
    INFO: BigQuery extract job completed: beam_job_6825f7721c99487bae2bb643012d69f6_bigqueryreadwriteit0testsqlreadjenkins0813115728361b09b-extract
    Aug 13, 2018 11:57:30 AM org.apache.beam.sdk.io.gcp.bigquery.BigQuerySourceBase split
    INFO: Extract job produced 1 files
    Aug 13, 2018 11:57:30 AM org.apache.beam.sdk.io.FileBasedSource createReader
    INFO: Matched 1 files for pattern gs://temp-storage-for-end-to-end-tests/BigQueryExtractTemp/6825f7721c99487bae2bb643012d69f6/000000000000.avro
    Aug 13, 2018 11:57:31 AM org.apache.beam.sdk.io.FileBasedSource getEstimatedSizeBytes
    INFO: Filepattern gs://temp-storage-for-end-to-end-tests/BigQueryExtractTemp/6825f7721c99487bae2bb643012d69f6/000000000000.avro matched 1 files with total size 738

org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryReadWriteIT > testInsertSelect STANDARD_ERROR
    Aug 13, 2018 11:57:31 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testInsertSelect_2018_08_13_11_57_31_662_2251103415000725556
    Aug 13, 2018 11:57:31 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$DatasetServiceImpl createTable
    INFO: Trying to create BigQuery table: apache-beam-testing:integ_test.BigQueryReadWriteIT_testInsertSelect_2018_08_13_11_57_31_887_3193178980064124757
    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    INSERT INTO `beam`.`ORDERS_BQ`
    (SELECT `ORDERS_IN_MEMORY`.`id` AS `id`, `ORDERS_IN_MEMORY`.`name` AS `name`, `ORDERS_IN_MEMORY`.`arr` AS `arr`
    FROM `beam`.`ORDERS_IN_MEMORY` AS `ORDERS_IN_MEMORY`)
    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    BeamIOSinkRel(table=[[beam, ORDERS_BQ]], operation=[INSERT], flattened=[true])
      LogicalProject(id=[$0], name=[$1], arr=[$2])
        BeamIOSourceRel(table=[[beam, ORDERS_IN_MEMORY]])

    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamIOSinkRel(table=[[beam, ORDERS_BQ]], operation=[INSERT], flattened=[true])
      BeamCalcRel(expr#0..2=[{inputs}], proj#0..2=[{exprs}])
        BeamIOSourceRel(table=[[beam, ORDERS_IN_MEMORY]])

    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.io.gcp.bigquery.BatchLoads$4 getTempFilePrefix
    INFO: Writing BigQuery temporary files to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f/ before loading them.
    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f/83322015-923d-4d86-bf6f-c368769e1f51.
    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f/d09f41c3-452e-47c7-8bda-aefb86b46766.
    Aug 13, 2018 11:57:32 AM org.apache.beam.sdk.io.gcp.bigquery.TableRowWriter <init>
    INFO: Opening TableRowWriter to gs://temp-storage-for-end-to-end-tests/BigQueryWriteTemp/beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f/d5c6e10e-a89c-4c8f-b79b-ddaf895a9267.
    Aug 13, 2018 11:57:33 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Loading 3 files into {datasetId=integ_test, projectId=apache-beam-testing, tableId=BigQueryReadWriteIT_testInsertSelect_2018_08_13_11_57_31_887_3193178980064124757} using job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0, location=US, projectId=apache-beam-testing}, attempt 1
    Aug 13, 2018 11:57:33 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl startJob
    INFO: Started BigQuery job: {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0, location=US, projectId=apache-beam-testing}.
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0
    Aug 13, 2018 11:57:33 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0, location=US, projectId=apache-beam-testing} started
    Aug 13, 2018 11:57:33 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: Still waiting for BigQuery job beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0, currently in status {"state":"RUNNING"}
    bq show -j --format=prettyjson --project_id=apache-beam-testing beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0
    Aug 13, 2018 11:57:35 AM org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl$JobServiceImpl pollJob
    INFO: BigQuery job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0, location=US, projectId=apache-beam-testing} completed in state DONE
    Aug 13, 2018 11:57:35 AM org.apache.beam.sdk.io.gcp.bigquery.WriteTables load
    INFO: Load job {jobId=beam_load_bigqueryreadwriteit0testinsertselectjenkins0813115732f69b8963_e8444f4d0d08492a8bef7911b9a59b3f_a0ff883d7743ec3daccd67b9703c4537_00001_00000-0, location=US, projectId=apache-beam-testing} succeeded. Statistics: {"creationTime":"1534161453508","endTime":"1534161454922","load":{"badRecords":"0","inputFileBytes":"126","inputFiles":"3","outputBytes":"69","outputRows":"3"},"startTime":"1534161453865"}

Gradle Test Executor 120 finished executing tests.
Gradle Test Executor 122 finished executing tests.

> Task :beam-sdks-java-extensions-sql:integrationTest

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testSelectsPayloadContent STANDARD_ERROR
    Aug 13, 2018 11:57:51 AM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription--8890081836908198994 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testUsesDlq STANDARD_ERROR
    Aug 13, 2018 11:57:56 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQL:
    SELECT `message`.`payload`.`id`, `message`.`payload`.`name`
    FROM `beam`.`message` AS `message`
    Aug 13, 2018 11:57:56 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: SQLPlan>
    LogicalProject(id=[$2], name=[$3])
      LogicalProject(event_timestamp=[$0], attributes=[$1], id=[$2.id], name=[$2.name])
        BeamIOSourceRel(table=[[beam, message]])

    Aug 13, 2018 11:57:56 AM org.apache.beam.sdk.extensions.sql.impl.BeamQueryPlanner convertToBeamRel
    INFO: BEAMPlan>
    BeamCalcRel(expr#0..2=[{inputs}], expr#3=[$t2.id], expr#4=[$t2.name], id=[$t3], name=[$t4])
      BeamIOSourceRel(table=[[beam, message]])

    Aug 13, 2018 11:58:00 AM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testUsesDlq-2018-08-13-11-57-56-244-events-626316813178105561_beam_-1172941878441060925 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testUsesDlq-2018-08-13-11-57-56-244-events-626316813178105561. Note this subscription WILL NOT be deleted when the pipeline terminates
    Aug 13, 2018 11:58:01 AM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testUsesDlq-2018-08-13-11-57-56-059-events-3509529891555216900_beam_-7653001277277239598 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testUsesDlq-2018-08-13-11-57-56-059-events-3509529891555216900. Note this subscription WILL NOT be deleted when the pipeline terminates
    Aug 13, 2018 11:58:16 AM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/start-subscription-8416247507066476569 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}
    Aug 13, 2018 11:58:40 AM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription-6074494511496792985 for signal: Status{code=DEADLINE_EXCEEDED, description=deadline exceeded after 14999446909ns, cause=null}
    Aug 13, 2018 11:58:55 AM org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal pollForResultForDuration
    WARNING: (Will retry) Error while polling projects/apache-beam-testing/subscriptions/result-subscription-6074494511496792985 for signal: Status{code=DEADLINE_EXCEEDED, description=Deadline expired before operation could complete., cause=null}

org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonIT > testSQLLimit STANDARD_ERROR
    Aug 13, 2018 11:59:09 AM org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource createRandomSubscription
    WARNING: Created subscription projects/apache-beam-testing/subscriptions/integ-test-PubsubJsonIT-testSQLLimit-2018-08-13-11-59-06-632-events--1505947905786808702_beam_-4762128899322970115 to topic projects/apache-beam-testing/topics/integ-test-PubsubJsonIT-testSQLLimit-2018-08-13-11-59-06-632-events--1505947905786808702. Note this subscription WILL NOT be deleted when the pipeline terminates

Gradle Test Executor 121 finished executing tests.

> Task :beam-sdks-java-extensions-sql:integrationTest
Finished generating test XML results (0.001 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.005 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/reports/tests/integrationTest>
Packing task ':beam-sdks-java-extensions-sql:integrationTest'
:beam-sdks-java-extensions-sql:integrationTest (Thread[Task worker for ':' Thread 3,5,main]) completed. Took 3 mins 0.384 secs.
:beam-sdks-java-extensions-sql:postCommit (Thread[Task worker for ':' Thread 3,5,main]) started.

> Task :beam-sdks-java-extensions-sql:postCommit
Skipping task ':beam-sdks-java-extensions-sql:postCommit' as it has no actions.
:beam-sdks-java-extensions-sql:postCommit (Thread[Task worker for ':' Thread 3,5,main]) completed. Took 0.0 secs.

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-extensions-google-cloud-platform-core:compileTestJava'.
> Compilation failed with exit code 1; see the compiler error output for details.

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':beam-sdks-java-io-elasticsearch-tests-2:compileTestJava'.
> Compilation failed with exit code 1; see the compiler error output for details.

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 15m 10s
702 actionable tasks: 701 executed, 1 from cache

Publishing build scan...
https://gradle.com/s/wnmf6lakefpm2

Stopped 1 worker daemon(s).
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

Jenkins build is back to normal : beam_PostCommit_Java_GradleBuild #1242

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/1242/display/redirect?page=changes>


Build failed in Jenkins: beam_PostCommit_Java_GradleBuild #1241

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/1241/display/redirect>

------------------------------------------
[...truncated 19.88 MB...]
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT > testReportFailures STANDARD_ERROR
    Aug 13, 2018 12:44:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2018-08-13_05_44_40-10650241323969448948
    Aug 13, 2018 12:44:41 PM org.apache.beam.runners.dataflow.TestDataflowRunner run
    INFO: Running Dataflow job 2018-08-13_05_44_40-10650241323969448948 with 0 expected assertions.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:40.452Z: Autoscaling is enabled for job 2018-08-13_05_44_40-10650241323969448948. The number of workers will be between 1 and 1000.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:40.536Z: Autoscaling was automatically enabled for job 2018-08-13_05_44_40-10650241323969448948.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:43.462Z: Checking required Cloud APIs are enabled.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:43.771Z: Checking permissions granted to controller Service Account.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:47.853Z: Worker configuration: n1-standard-1 in us-central1-b.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:48.370Z: Expanding CoGroupByKey operations into optimizable parts.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:48.623Z: Expanding GroupByKey operations into optimizable parts.
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:48.668Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:48.940Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:48.993Z: Elided trivial flatten 
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.038Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/Wait/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Create seed/Read(CreateSource)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.084Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Read information schema into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/Wait/Map
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.131Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.168Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow) into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.211Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Read information schema
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.254Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.301Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.341Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey) into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.383Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.430Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.476Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.519Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.557Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/ParDo(CollectWindows) into SpannerIO.Write/To mutation group
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.594Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Extract keys into SpannerIO.Write/Write mutations to Cloud Spanner/Serialize mutations
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.632Z: Fusing consumer ParDo(GenerateMutations) into GenerateSequence/Read(BoundedCountingSource)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.670Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.707Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Batch mutations together into SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/GroupByWindow
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.742Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike) into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.787Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.834Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow) into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Extract
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.875Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Reify
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.922Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Extract into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:49.965Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Partial
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.004Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Partial into SpannerIO.Write/Write mutations to Cloud Spanner/Extract keys
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.047Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Reify
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.094Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.139Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/View.AsList/ParDo(ToIsmRecordForGlobalWindow) into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Flatten.Iterables/FlattenIterables/FlatMap
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.184Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.222Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Flatten.Iterables/FlattenIterables/FlatMap into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Values/Values/Map
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.257Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Values/Values/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Extract
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.293Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForKey) into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.330Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Extract into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.377Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/WithKeys/AddKeys/Map into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/ParDo(CollectWindows)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.425Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.471Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Write mutations to Spanner into SpannerIO.Write/Write mutations to Cloud Spanner/Batch mutations together
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.509Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Partial
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.546Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Partial into SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/WithKeys/AddKeys/Map
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.594Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Reify into SpannerIO.Write/Write mutations to Cloud Spanner/Partition input
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.634Z: Fusing consumer SpannerIO.Write/To mutation group into ParDo(GenerateMutations)
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.681Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/GroupByWindow into SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.720Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Write into SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Reify
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:50.761Z: Fusing consumer SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForSize) into SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Read
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.201Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.239Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.270Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.296Z: Starting 1 workers in us-central1-b...
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.309Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.347Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.375Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.413Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.450Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Create
    Aug 13, 2018 12:44:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:44:51.808Z: Executing operation GenerateSequence/Read(BoundedCountingSource)+ParDo(GenerateMutations)+SpannerIO.Write/To mutation group+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/ParDo(CollectWindows)+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/WithKeys/AddKeys/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Partial+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Write
    Aug 13, 2018 12:45:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:45:01.326Z: Autoscaling: Raised the number of workers to 0 based on the rate of progress in the currently running step(s).
    Aug 13, 2018 12:45:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:45:11.869Z: Autoscaling: Raised the number of workers to 1 based on the rate of progress in the currently running step(s).
    Aug 13, 2018 12:45:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:45:11.911Z: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job.
    Aug 13, 2018 12:45:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:45:41.438Z: Workers have started successfully.
    Aug 13, 2018 12:45:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:45:41.468Z: Workers have started successfully.
    Aug 13, 2018 12:46:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:19.722Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Close
    Aug 13, 2018 12:46:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:19.827Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/GroupByKey/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Combine.perKey(SampleAny)/Combine.GroupedValues/Extract+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Combine.globally(SampleAny)/Values/Values/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/Sample.Any/Flatten.Iterables/FlattenIterables/FlatMap+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/View.AsList/ParDo(ToIsmRecordForGlobalWindow)
    Aug 13, 2018 12:46:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:27.707Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/To wait view 0/View.AsList/CreateDataflowView
    Aug 13, 2018 12:46:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:27.919Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Create seed/Read(CreateSource)+SpannerIO.Write/Write mutations to Cloud Spanner/Wait.OnSignal/Wait/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Read information schema+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write
    Aug 13, 2018 12:46:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:35.864Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Close
    Aug 13, 2018 12:46:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:35.947Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
    Aug 13, 2018 12:46:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:40.533Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
    Aug 13, 2018 12:46:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:40.607Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow)
    Aug 13, 2018 12:46:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:45.854Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Schema View/Combine.GloballyAsSingletonView/CreateDataflowView
    Aug 13, 2018 12:46:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:46.024Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Serialize mutations+SpannerIO.Write/Write mutations to Cloud Spanner/Extract keys+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Partial+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Write
    Aug 13, 2018 12:46:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:53.877Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Close
    Aug 13, 2018 12:46:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:53.956Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/GroupByKey/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud Spanner/Sample keys/Combine.GroupedValues/Extract+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow)+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
    Aug 13, 2018 12:46:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:56.373Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
    Aug 13, 2018 12:46:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:46:56.467Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParMultiDo(ToIsmRecordForMapLike)+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Write+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Write
    Aug 13, 2018 12:47:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:02.913Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Close
    Aug 13, 2018 12:47:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:02.948Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Close
    Aug 13, 2018 12:47:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:02.996Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForSize)
    Aug 13, 2018 12:47:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:03.041Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/ParDo(ToIsmMetadataRecordForKey)
    Aug 13, 2018 12:47:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:09.421Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/Flatten.PCollections
    Aug 13, 2018 12:47:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:09.634Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Keys sample as view/CreateDataflowView
    Aug 13, 2018 12:47:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:09.816Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Partition input+SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Reify+SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Write
    Aug 13, 2018 12:47:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:16.013Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Close
    Aug 13, 2018 12:47:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:16.103Z: Executing operation SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Group by partition/GroupByWindow+SpannerIO.Write/Write mutations to Cloud Spanner/Batch mutations together+SpannerIO.Write/Write mutations to Cloud Spanner/Write mutations to Spanner
    Aug 13, 2018 12:47:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:21.454Z: Cleaning up.
    Aug 13, 2018 12:47:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:47:21.725Z: Stopping worker pool...
    Aug 13, 2018 12:48:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:48:47.296Z: Autoscaling: Resized worker pool from 1 to 0.
    Aug 13, 2018 12:48:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:48:47.335Z: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job.
    Aug 13, 2018 12:48:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-08-13T12:48:47.394Z: Worker pool stopped.
    Aug 13, 2018 12:48:55 PM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-08-13_05_44_40-10650241323969448948 finished with status DONE.
    Aug 13, 2018 12:48:55 PM org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    INFO: Success result for Dataflow job 2018-08-13_05_44_40-10650241323969448948. Found 0 success, 0 failures out of 0 expected assertions.
    Aug 13, 2018 12:48:55 PM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-08-13_05_44_40-10650241323969448948 finished with status DONE.

Gradle Test Executor 146 finished executing tests.

> Task :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest
Finished generating test XML results (0.01 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.014 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest>
Packing task ':beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest'
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest (Thread[Task worker for ':' Thread 4,5,main]) completed. Took 32 mins 12.477 secs.
:beam-runners-google-cloud-dataflow-java:buildDependents (Thread[Task worker for ':' Thread 4,5,main]) started.

> Task :beam-runners-google-cloud-dataflow-java:buildDependents
Caching disabled for task ':beam-runners-google-cloud-dataflow-java:buildDependents': Caching has not been enabled for the task
Task ':beam-runners-google-cloud-dataflow-java:buildDependents' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
:beam-runners-google-cloud-dataflow-java:buildDependents (Thread[Task worker for ':' Thread 4,5,main]) completed. Took 0.0 secs.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Daemon worker,5,main]) started.

> Task :beam-runners-google-cloud-dataflow-java:postCommit
Skipping task ':beam-runners-google-cloud-dataflow-java:postCommit' as it has no actions.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Daemon worker,5,main]) completed. Took 0.0 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':beam-sdks-java-nexmark:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/nexmark/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 48m 36s
751 actionable tasks: 746 executed, 5 from cache

Publishing build scan...
https://gradle.com/s/qvlhzb74rzjmo

Build cache (/home/jenkins/.gradle/caches/build-cache-1) remove files older than Mon Aug 06 12:00:31 UTC 2018.
Build cache (/home/jenkins/.gradle/caches/build-cache-1) removing 1001 cache entries (61 MB reclaimed).
Build cache (/home/jenkins/.gradle/caches/build-cache-1) cleaned up in 0.289 secs.
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure