You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/05/16 12:46:04 UTC

Build failed in Jenkins: beam_PostCommit_XVR_Spark #846

See <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/846/display/redirect?page=changes>

Changes:

[github] Merge pull request #11725: [BEAM-10015] Fix output timestamp on dataflow


------------------------------------------
[...truncated 1022.50 KB...]
    self._block(timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/local/lib/python2.7/site-packages/grpc/_utilities.py",> line 86, in _block
    raise grpc.FutureTimeoutError()
FutureTimeoutError: 
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.22.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.22.0-SNAPSHOT.jar'> '56203']
root: DEBUG: Waiting for grpc channel to be ready at localhost:56203.
apache_beam.utils.subprocess_server: INFO: Starting expansion service at localhost:56203
apache_beam.utils.subprocess_server: INFO: 	beam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/598446861@deb6432
apache_beam.utils.subprocess_server: INFO: 	beam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/598446861@28ba21f3
root: DEBUG: Waiting for grpc channel to be ready at localhost:56203.
root: DEBUG: Waiting for grpc channel to be ready at localhost:56203.
root: DEBUG: Waiting for grpc channel to be ready at localhost:56203.
root: DEBUG: Waiting for grpc channel to be ready at localhost:56203.
root: DEBUG: Waiting for grpc channel to be ready at localhost:56203.
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:42:34 PM org.apache.beam.sdk.expansion.service.ExpansionService expand
apache_beam.utils.subprocess_server: INFO: INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:42:36 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: SQL:
apache_beam.utils.subprocess_server: INFO: SELECT CAST(1 AS INTEGER) AS `int`, CAST('foo' AS VARCHAR) AS `str`, CAST(3.14 AS DOUBLE) AS `flt`
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:42:37 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: SQLPlan>
apache_beam.utils.subprocess_server: INFO: LogicalProject(int=[1], str=['foo':VARCHAR], flt=[3.14:DOUBLE])
apache_beam.utils.subprocess_server: INFO:   LogicalValues(tuples=[[{ 0 }]])
apache_beam.utils.subprocess_server: INFO: 
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:42:37 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: BEAMPlan>
apache_beam.utils.subprocess_server: INFO: BeamCalcRel(expr#0=[{inputs}], expr#1=[1], expr#2=['foo':VARCHAR], expr#3=[3.14:DOUBLE], int=[$t1], str=[$t2], flt=[$t3])
apache_beam.utils.subprocess_server: INFO:   BeamValuesRel(tuples=[[{ 0 }]])
apache_beam.utils.subprocess_server: INFO: 
root: DEBUG: Sending SIGINT to job_server
root: WARNING: Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
root: INFO: Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
root: WARNING: Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
root: INFO: Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f5ee81d6320> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 16 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['external_3SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_15/Create.Values/Read(CreateSource)/Impulse\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_15/Create.Values/Read(CreateSource)/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_3SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_15/Create.Values/Read(CreateSource)/MapElements/Map/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_15/Create.Values/Read(CreateSource)/MapElements/Map/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_3SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_15/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper)\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_15/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_3SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_19/ParDo(Calc)/ParMultiDo(Calc)\n  SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_19/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_5\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2713>)_6\n  assert_that/Create/FlatMap(<lambda at core.py:2713>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_8\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_9\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_10\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_12\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_13\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_14\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_15\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_19\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_20\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_21\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_project (apache_beam.transforms.sql_test.SqlTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/transforms/sql_test.py",> line 84, in test_project
    assert_that(out, equal_to([(1, 3.14)]))
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 546, in __exit__
    self.run().wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 112, in run
    False if self.not_use_test_runner_api else test_runner_api))
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 525, in run
    return self.runner.run_pipeline(self, self._options)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 424, in run_pipeline
    job_service_handle = self.create_job_service(options)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 331, in create_job_service
    return self.create_job_service_handle(server.start(), options)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/job_server.py",> line 56, in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/local/lib/python2.7/site-packages/grpc/_utilities.py",> line 140, in result
    self._block(timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/local/lib/python2.7/site-packages/grpc/_utilities.py",> line 86, in _block
    raise grpc.FutureTimeoutError()
FutureTimeoutError: 
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.22.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.22.0-SNAPSHOT.jar'> '44559']
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
apache_beam.utils.subprocess_server: INFO: Starting expansion service at localhost:44559
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
apache_beam.utils.subprocess_server: INFO: 	beam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/598446861@deb6432
apache_beam.utils.subprocess_server: INFO: 	beam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/598446861@28ba21f3
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
root: DEBUG: Waiting for grpc channel to be ready at localhost:44559.
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:43:44 PM org.apache.beam.sdk.expansion.service.ExpansionService expand
apache_beam.utils.subprocess_server: INFO: INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:43:47 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: SQL:
apache_beam.utils.subprocess_server: INFO: SELECT `PCOLLECTION`.`int`, `PCOLLECTION`.`flt`
apache_beam.utils.subprocess_server: INFO: FROM `beam`.`PCOLLECTION` AS `PCOLLECTION`
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:43:48 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: SQLPlan>
apache_beam.utils.subprocess_server: INFO: LogicalProject(int=[$0], flt=[$2])
apache_beam.utils.subprocess_server: INFO:   BeamIOSourceRel(table=[[beam, PCOLLECTION]])
apache_beam.utils.subprocess_server: INFO: 
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:43:48 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: BEAMPlan>
apache_beam.utils.subprocess_server: INFO: BeamCalcRel(expr#0..2=[{inputs}], int=[$t0], flt=[$t2])
apache_beam.utils.subprocess_server: INFO:   BeamIOSourceRel(table=[[beam, PCOLLECTION]])
apache_beam.utils.subprocess_server: INFO: 
root: DEBUG: Sending SIGINT to job_server
root: WARNING: Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
root: INFO: Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
root: WARNING: Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
root: INFO: Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f5ee81d6320> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 17 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_Create/Impulse_3\n  Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2713>)_4\n  Create/FlatMap(<lambda at core.py:2713>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/Map(decode)_6\n  Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_4SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_2/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_2/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_4SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_18/ParDo(Calc)/ParMultiDo(Calc)\n  SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_18/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_10\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2713>)_11\n  assert_that/Create/FlatMap(<lambda at core.py:2713>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_13\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_14\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_15\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_17\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_18\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_19\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_20\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_24\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_25\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_26\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_tagged_join (apache_beam.transforms.sql_test.SqlTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/transforms/sql_test.py",> line 139, in test_tagged_join
    assert_that(out, equal_to([(1, "a"), (26, "z"), (1, "a")]))
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 546, in __exit__
    self.run().wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 112, in run
    False if self.not_use_test_runner_api else test_runner_api))
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 525, in run
    return self.runner.run_pipeline(self, self._options)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 424, in run_pipeline
    job_service_handle = self.create_job_service(options)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 331, in create_job_service
    return self.create_job_service_handle(server.start(), options)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/job_server.py",> line 56, in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/local/lib/python2.7/site-packages/grpc/_utilities.py",> line 140, in result
    self._block(timeout)
  File "<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/local/lib/python2.7/site-packages/grpc/_utilities.py",> line 86, in _block
    raise grpc.FutureTimeoutError()
FutureTimeoutError: 
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.22.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.22.0-SNAPSHOT.jar'> '51593']
root: DEBUG: Waiting for grpc channel to be ready at localhost:51593.
apache_beam.utils.subprocess_server: INFO: Starting expansion service at localhost:51593
root: DEBUG: Waiting for grpc channel to be ready at localhost:51593.
apache_beam.utils.subprocess_server: INFO: 	beam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/598446861@deb6432
apache_beam.utils.subprocess_server: INFO: 	beam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/598446861@28ba21f3
root: DEBUG: Waiting for grpc channel to be ready at localhost:51593.
root: DEBUG: Waiting for grpc channel to be ready at localhost:51593.
root: DEBUG: Waiting for grpc channel to be ready at localhost:51593.
root: DEBUG: Waiting for grpc channel to be ready at localhost:51593.
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:44:53 PM org.apache.beam.sdk.expansion.service.ExpansionService expand
apache_beam.utils.subprocess_server: INFO: INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:44:55 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: SQL:
apache_beam.utils.subprocess_server: INFO: SELECT `simple`.`int` AS `int`, `enrich`.`metadata` AS `metadata`
apache_beam.utils.subprocess_server: INFO: FROM `beam`.`simple` AS `simple`
apache_beam.utils.subprocess_server: INFO: INNER JOIN `beam`.`enrich` AS `enrich` ON `simple`.`int` = `enrich`.`int`
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:44:55 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: SQLPlan>
apache_beam.utils.subprocess_server: INFO: LogicalProject(int=[$0], metadata=[$4])
apache_beam.utils.subprocess_server: INFO:   LogicalJoin(condition=[=($0, $3)], joinType=[inner])
apache_beam.utils.subprocess_server: INFO:     BeamIOSourceRel(table=[[beam, simple]])
apache_beam.utils.subprocess_server: INFO:     BeamIOSourceRel(table=[[beam, enrich]])
apache_beam.utils.subprocess_server: INFO: 
apache_beam.utils.subprocess_server: INFO: May 16, 2020 12:44:56 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel
apache_beam.utils.subprocess_server: INFO: INFO: BEAMPlan>
apache_beam.utils.subprocess_server: INFO: BeamCalcRel(expr#0..4=[{inputs}], int=[$t2], metadata=[$t1])
apache_beam.utils.subprocess_server: INFO:   BeamCoGBKJoinRel(condition=[=($2, $0)], joinType=[inner])
apache_beam.utils.subprocess_server: INFO:     BeamIOSourceRel(table=[[beam, enrich]])
apache_beam.utils.subprocess_server: INFO:     BeamIOSourceRel(table=[[beam, simple]])
apache_beam.utils.subprocess_server: INFO: 
root: DEBUG: Sending SIGINT to job_server
root: WARNING: Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
root: INFO: Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
root: WARNING: Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
root: INFO: Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f5ee81d6320> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 42 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_Create enrich/Impulse_3\n  Create enrich/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/FlatMap(<lambda at core.py:2713>)_4\n  Create enrich/FlatMap(<lambda at core.py:2713>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n  Create enrich/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n  Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n  Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n  Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/RemoveRandomKeys_15\n  Create enrich/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/Map(decode)_16\n  Create enrich/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/Impulse_18\n  Create simple/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/FlatMap(<lambda at core.py:2713>)_19\n  Create simple/FlatMap(<lambda at core.py:2713>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/AddRandomKeys_22\n  Create simple/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_24\n  Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_25\n  Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_29\n  Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/RemoveRandomKeys_30\n  Create simple/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/Map(decode)_31\n  Create simple/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_6/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_6/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_5/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_5/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/left_TimestampCombiner/Flatten.PCollections\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/left_TimestampCombiner/Flatten.PCollections:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/right_TimestampCombiner/Flatten.PCollections\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/right_TimestampCombiner/Flatten.PCollections:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeylhs/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeylhs/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeyrhs/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeyrhs/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/Flatten\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/GBK\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/GBK:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/ParDo(ConvertCoGbkResult)/ParMultiDo(ConvertCoGbkResult)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/ParDo(ConvertCoGbkResult)/ParMultiDo(ConvertCoGbkResult):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Select.Fields/ParDo(Select)/ParMultiDo(Select)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Select.Fields/ParDo(Select)/ParMultiDo(Select):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_5SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_96/ParDo(Calc)/ParMultiDo(Calc)\n  SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_96/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_35\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2713>)_36\n  assert_that/Create/FlatMap(<lambda at core.py:2713>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_38\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_39\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_40\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_42\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_43\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_44\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_45\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_49\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_50\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_51\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-xlangSqlValidateRunner.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 5 tests in 345.923s

FAILED (errors=5)

> Task :runners:spark:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:job-server:sparkJobServerCleanup

FAILURE: Build completed with 5 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:job-server:validatesCrossLanguageRunnerJavaUsingJava'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingJava/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:job-server:validatesCrossLanguageRunnerJavaUsingPython'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingPython/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

5: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 40m 8s
131 actionable tasks: 100 executed, 29 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/4paj262vh7tzw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_XVR_Spark #847

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/847/display/redirect>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org