You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/04/21 18:47:44 UTC

Build failed in Jenkins: beam_PostCommit_XVR_Spark #2206

See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/2206/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-12009] Copy CalcRelSplitter.

[Kyle Weaver] [BEAM-12009] Implement Calc splitting rule.

[Kyle Weaver] [BEAM-12009] Use different method signature to avoid spurious null check

[Kyle Weaver] [BEAM-12009] Move CalcRelSplitter to rel subpackage.

[Kyle Weaver] [BEAM-12009] Reduce code duplication between rules by re-implementing

[Kyle Weaver] [BEAM-12009] Only match in BeamCalcSplittingRule if ≥ 1 of its component

[anup.d] BEAM-12166:Beam Sql - Combine Accumulator return Map fails with class

[suztomo] [BEAM-8357] Upgrading auto-value to 1.8 from 1.7.4

[Kenneth Knowles] Build source release zip from RC tag

[aromanenko.dev] [BEAM-12197] TPC-DS: Fix SQL-queries syntax


------------------------------------------
[...truncated 1006.95 KB...]
    state = result.wait_until_finish()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 602, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline BeamApp-jenkins-0421184701-28837d23_b6a424c4-85d4-4948-8966-705d1acaed2a failed in state FAILED: java.lang.NullPointerException
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar'> '35769']
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
apache_beam.utils.subprocess_server: INFO: b'Starting expansion service at localhost:35769'
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:46:26 PM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
apache_beam.utils.subprocess_server: INFO: b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5680a178'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5fdef03a'
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35769.
root: WARNING: Waiting for grpc channel to be ready at localhost:35769.
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:46:31 PM org.apache.beam.sdk.expansion.service.ExpansionService expand'
apache_beam.utils.subprocess_server: INFO: b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:46:35 PM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
apache_beam.utils.subprocess_server: INFO: b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:46:44 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
apache_beam.utils.subprocess_server: INFO: b'INFO: SQL:'
apache_beam.utils.subprocess_server: INFO: b'SELECT COUNT(*) AS `count`'
apache_beam.utils.subprocess_server: INFO: b'FROM `beam`.`PCOLLECTION` AS `PCOLLECTION`'
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:46:45 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
apache_beam.utils.subprocess_server: INFO: b'INFO: SQLPlan>'
apache_beam.utils.subprocess_server: INFO: b'LogicalAggregate(group=[{}], count=[COUNT()])'
apache_beam.utils.subprocess_server: INFO: b'  LogicalProject($f0=[0])'
apache_beam.utils.subprocess_server: INFO: b'    BeamIOSourceRel(table=[[beam, PCOLLECTION]])'
apache_beam.utils.subprocess_server: INFO: b''
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:46:46 PM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
apache_beam.utils.subprocess_server: INFO: b'INFO: BEAMPlan>'
apache_beam.utils.subprocess_server: INFO: b'BeamAggregationRel(group=[{}], count=[COUNT()])'
apache_beam.utils.subprocess_server: INFO: b'  BeamIOSourceRel(table=[[beam, PCOLLECTION]])'
apache_beam.utils.subprocess_server: INFO: b''
root: DEBUG: Sending SIGINT to job_server
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: No image given, using default Python SDK image
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: Python SDK container image set to "apache/beam_python3.6_sdk:2.30.0.dev" for Docker environment
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f6e03098d90> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 27 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_Create-Impulse_3\n  Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-FlatMap-lambda-at-core-py-2955-_4\n  Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-AddRandomKeys_7\n  Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-Map-reify_timestamps-_9\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-GroupByKey_10\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-FlatMap-restore_timestamps-_11\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-RemoveRandomKeys_12\n  Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-Map-decode-_13\n  Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map-lambda-at-sql_test-py-174-_14\n  Map(<lambda at sql_test.py:174>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_WindowInto-WindowIntoFn-_15\n  WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-sele\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/selectKeys/AddKeys/Map/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-Grou\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-Combine-Pa\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/Combine/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToRow-ParM\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToRow/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-mergeRecord-ParMultiDo-Anonymous-\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/mergeRecord/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_19\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_20\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_22\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_23\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_24\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_26\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_27\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_28\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_29\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_30\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_31\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_32\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function sort_stages at 0x7f6e03099510> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 27 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_Create-Impulse_3\n  Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-FlatMap-lambda-at-core-py-2955-_4\n  Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-AddRandomKeys_7\n  Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-Map-reify_timestamps-_9\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-GroupByKey_10\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-FlatMap-restore_timestamps-_11\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-RemoveRandomKeys_12\n  Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-Map-decode-_13\n  Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map-lambda-at-sql_test-py-174-_14\n  Map(<lambda at sql_test.py:174>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_WindowInto-WindowIntoFn-_15\n  WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-sele\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/selectKeys/AddKeys/Map/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-Grou\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-Combine-Pa\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/Combine/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToRow-ParM\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToRow/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-mergeRecord-ParMultiDo-Anonymous-\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/mergeRecord/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_19\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_20\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_22\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_23\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_24\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_26\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_27\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_28\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_29\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_30\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_31\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_32\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'experiments' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_name' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'runner' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'temp_location' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'dataflow_kms_key' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'enable_streaming_engine' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'project' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_region' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'streaming' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_cache_millis' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_options' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_endpoint' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'output_executable_path' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'sdk_worker_parallelism' was already added
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: java.lang.NullPointerException
	at org.apache.beam.runners.spark.SparkPipelineRunner.run(SparkPipelineRunner.java:120)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: java.lang.NullPointerException
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_zetasql_generate_data (apache_beam.transforms.sql_test.SqlTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/transforms/sql_test.py",> line 164, in test_zetasql_generate_data
    assert_that(out, equal_to([(1, "foo", 3.14)]))
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 582, in __exit__
    self.result = self.run()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 112, in run
    state = result.wait_until_finish()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 602, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline BeamApp-jenkins-0421184736-4efe5d34_c58c6d2f-fa5c-4586-a43e-96e038ebe5b9 failed in state FAILED: java.lang.NullPointerException
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar'> '35563']
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
apache_beam.utils.subprocess_server: INFO: b'Starting expansion service at localhost:35563'
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:47:03 PM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
apache_beam.utils.subprocess_server: INFO: b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5680a178'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5fdef03a'
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: DEBUG: Waiting for grpc channel to be ready at localhost:35563.
root: WARNING: Waiting for grpc channel to be ready at localhost:35563.
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:47:07 PM org.apache.beam.sdk.expansion.service.ExpansionService expand'
apache_beam.utils.subprocess_server: INFO: b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:47:12 PM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
apache_beam.utils.subprocess_server: INFO: b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
apache_beam.utils.subprocess_server: INFO: b'Apr 21, 2021 6:47:25 PM org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner convertToBeamRelInternal'
apache_beam.utils.subprocess_server: INFO: b'INFO: BEAMPlan>'
apache_beam.utils.subprocess_server: INFO: b"BeamZetaSqlCalcRel(expr#0=[{inputs}], expr#1=[1:BIGINT], expr#2=['foo':VARCHAR], expr#3=[3.1400000000000001243E0:DOUBLE], int=[$t1], str=[$t2], flt=[$t3])"
apache_beam.utils.subprocess_server: INFO: b'  BeamValuesRel(tuples=[[{ 0 }]])'
apache_beam.utils.subprocess_server: INFO: b''
root: DEBUG: Sending SIGINT to job_server
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: No image given, using default Python SDK image
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: Python SDK container image set to "apache/beam_python3.6_sdk:2.30.0.dev" for Docker environment
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f6e03098d90> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 16 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--Impulse\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Outp\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Boun\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamZetaSqlCalcRel_17-ParDo-Calc--ParMultiDo-Calc-\n  SqlTransform(beam:external:java:sql:v1)/BeamZetaSqlCalcRel_17/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_5\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_6\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_8\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_9\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_10\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_12\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_13\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_14\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_15\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_16\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_17\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_18\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function sort_stages at 0x7f6e03099510> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 16 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--Impulse\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Outp\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Boun\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamZetaSqlCalcRel_17-ParDo-Calc--ParMultiDo-Calc-\n  SqlTransform(beam:external:java:sql:v1)/BeamZetaSqlCalcRel_17/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_5\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_6\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_8\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_9\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_10\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_12\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_13\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_14\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_15\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_16\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_17\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_18\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'experiments' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_name' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'runner' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'temp_location' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'dataflow_kms_key' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'enable_streaming_engine' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'project' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_region' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'streaming' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_cache_millis' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_options' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_endpoint' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'output_executable_path' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'sdk_worker_parallelism' was already added
root: DEBUG: java.lang.NullPointerException
	at org.apache.beam.runners.spark.SparkPipelineRunner.run(SparkPipelineRunner.java:120)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: java.lang.NullPointerException
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-xlangSqlValidateRunner.xml
----------------------------------------------------------------------
XML: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 9 tests in 290.484s

FAILED (errors=9)

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerCleanup
Stopping expansion service pid: 23446.
Stopping expansion service pid: 23451.

> Task :runners:spark:2:job-server:sparkJobServerCleanup
Stopping job server pid: 24992.

FAILURE: Build completed with 3 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 41m 19s
182 actionable tasks: 135 executed, 43 from cache, 4 up-to-date
Gradle was unable to watch the file system for changes. The inotify watches limit is too low.

Publishing build scan...
https://gradle.com/s/wphit6juwgouy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_XVR_Spark #2209

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/2209/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark #2208

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/2208/display/redirect?page=changes>

Changes:

[noreply] [BEAM-12191] Add a test for python template generation with upload_graph

[noreply] [BEAM-7372] fix wrong usage of with_traceback (#14566)

[noreply] [BEAM-7372] cleanup codes for py2 from apache_beam/transforms (#14544)

[noreply] [BEAM-2085] Fixups for Python resource hints. (#14605)


------------------------------------------
[...truncated 93.65 KB...]
  Using cached google_cloud_pubsub-1.7.0-py2.py3-none-any.whl (144 kB)
Collecting google-cloud-bigquery<3,>=1.6.0
  Using cached google_cloud_bigquery-2.13.1-py2.py3-none-any.whl (216 kB)
Collecting google-cloud-core<2,>=0.28.1
  Using cached google_cloud_core-1.6.0-py2.py3-none-any.whl (28 kB)
Collecting google-cloud-bigtable<2,>=0.31.1
  Using cached google_cloud_bigtable-1.7.0-py2.py3-none-any.whl (267 kB)
Collecting google-cloud-spanner<2,>=1.13.0
  Using cached google_cloud_spanner-1.19.1-py2.py3-none-any.whl (255 kB)
Collecting grpcio-gcp<1,>=0.2.2
  Using cached grpcio_gcp-0.2.2-py2.py3-none-any.whl (9.4 kB)
Collecting google-cloud-dlp<2,>=0.12.0
  Using cached google_cloud_dlp-1.0.0-py2.py3-none-any.whl (169 kB)
Collecting google-cloud-language<2,>=1.3.0
  Using cached google_cloud_language-1.3.0-py2.py3-none-any.whl (83 kB)
Collecting google-cloud-videointelligence<2,>=1.8.0
  Using cached google_cloud_videointelligence-1.16.1-py2.py3-none-any.whl (183 kB)
Collecting google-cloud-vision<2,>=0.38.0
  Using cached google_cloud_vision-1.0.0-py2.py3-none-any.whl (435 kB)
Collecting freezegun>=0.3.12
  Using cached freezegun-1.1.0-py2.py3-none-any.whl (16 kB)
Collecting mock<3.0.0,>=1.0.1
  Using cached mock-2.0.0-py2.py3-none-any.whl (56 kB)
Collecting nose>=1.3.7
  Using cached nose-1.3.7-py3-none-any.whl (154 kB)
Collecting nose_xunitmp>=0.4.1
  Using cached nose_xunitmp-0.4.1-py3-none-any.whl
Collecting pandas<1.3.0,>=1.0
  Using cached pandas-1.1.5-cp36-cp36m-manylinux1_x86_64.whl (9.5 MB)
Collecting parameterized<0.8.0,>=0.7.1
  Using cached parameterized-0.7.5-py2.py3-none-any.whl (17 kB)
Collecting pyhamcrest!=1.10.0,<2.0.0,>=1.9
  Using cached PyHamcrest-1.10.1-py3-none-any.whl (48 kB)
Collecting pyyaml<6.0.0,>=3.12
  Using cached PyYAML-5.4.1-cp36-cp36m-manylinux1_x86_64.whl (640 kB)
Collecting requests_mock<2.0,>=1.7
  Using cached requests_mock-1.8.0-py2.py3-none-any.whl (23 kB)
Collecting tenacity<6.0,>=5.0.2
  Using cached tenacity-5.1.5-py2.py3-none-any.whl (34 kB)
Collecting pytest<5.0,>=4.4.0
  Using cached pytest-4.6.11-py2.py3-none-any.whl (231 kB)
Collecting pytest-xdist<2,>=1.29.0
  Using cached pytest_xdist-1.34.0-py2.py3-none-any.whl (36 kB)
Collecting pytest-timeout<2,>=1.3.3
  Using cached pytest_timeout-1.4.2-py2.py3-none-any.whl (10 kB)
Collecting sqlalchemy<2.0,>=1.3
  Downloading SQLAlchemy-1.4.11-cp36-cp36m-manylinux2014_x86_64.whl (1.5 MB)
Collecting psycopg2-binary<3.0.0,>=2.8.5
  Using cached psycopg2_binary-2.8.6-cp36-cp36m-manylinux1_x86_64.whl (3.0 MB)
Collecting testcontainers<4.0.0,>=3.0.3
  Using cached testcontainers-3.4.0-py2.py3-none-any.whl (31 kB)
Requirement already satisfied: six>=1.11.0 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from azure-core>=1.7.0->apache-beam==2.30.0.dev0) (1.15.0)

> Task :sdks:go:resolveBuildDependencies
Resolving golang.org/x/net: commit='2fb46b16b8dda405028c50f7c7f0f9dd1fa6bfb1', urls=[https://go.googlesource.com/net]
Resolving golang.org/x/oauth2: commit='5d25da1a8d43b66f2898c444f899c7bcfd6a407e', urls=[https://go.googlesource.com/oauth2]
Resolving golang.org/x/sync: commit='fd80eb99c8f653c847d294a001bdf2a3a6f768f5', urls=[https://go.googlesource.com/sync]
Resolving golang.org/x/sys: commit='fde4db37ae7ad8191b03d30d27f258b5291ae4e3', urls=[https://go.googlesource.com/sys]

> Task :sdks:python:installGcpTest
Collecting cryptography>=2.1.4
  Using cached cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl (3.2 MB)
Collecting msrest>=0.6.18
  Using cached msrest-0.6.21-py2.py3-none-any.whl (85 kB)
Collecting botocore<1.21.0,>=1.20.55
  Using cached botocore-1.20.55-py2.py3-none-any.whl (7.4 MB)
Collecting s3transfer<0.5.0,>=0.4.0
  Using cached s3transfer-0.4.1-py2.py3-none-any.whl (79 kB)
Collecting jmespath<1.0.0,>=0.7.1
  Using cached jmespath-0.10.0-py2.py3-none-any.whl (24 kB)
Collecting urllib3<1.27,>=1.25.4
  Using cached urllib3-1.26.4-py2.py3-none-any.whl (153 kB)
Collecting cffi>=1.12
  Using cached cffi-1.14.5-cp36-cp36m-manylinux1_x86_64.whl (401 kB)
Collecting pycparser
  Using cached pycparser-2.20-py2.py3-none-any.whl (112 kB)
Collecting fasteners>=0.14
  Using cached fasteners-0.16-py2.py3-none-any.whl (28 kB)
Collecting pyasn1-modules>=0.2.1
  Using cached pyasn1_modules-0.2.8-py2.py3-none-any.whl (155 kB)
Requirement already satisfied: setuptools>=40.3.0 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from google-auth<2,>=1.18.0->apache-beam==2.30.0.dev0) (56.0.0)
Collecting rsa<5,>=3.1.4
  Using cached rsa-4.7.2-py3-none-any.whl (34 kB)
Collecting google-resumable-media<2.0dev,>=0.6.0
  Using cached google_resumable_media-1.2.0-py2.py3-none-any.whl (75 kB)
Collecting proto-plus>=1.10.0
  Using cached proto_plus-1.18.1-py3-none-any.whl (42 kB)
Collecting google-api-core[grpc]<2.0.0dev,>=1.23.0
  Using cached google_api_core-1.26.3-py2.py3-none-any.whl (93 kB)
Requirement already satisfied: packaging>=14.3 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from google-cloud-bigquery<3,>=1.6.0->apache-beam==2.30.0.dev0) (20.9)
Collecting googleapis-common-protos<2.0dev,>=1.6.0
  Using cached googleapis_common_protos-1.53.0-py2.py3-none-any.whl (198 kB)
Collecting grpc-google-iam-v1<0.13dev,>=0.12.3
  Using cached grpc_google_iam_v1-0.12.3-py3-none-any.whl
Collecting google-crc32c<2.0dev,>=1.0
  Using cached google_crc32c-1.1.2-cp36-cp36m-manylinux2014_x86_64.whl (38 kB)
Collecting docopt
  Using cached docopt-0.6.2-py2.py3-none-any.whl
Requirement already satisfied: pyparsing<3,>=2.4.2 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from httplib2<0.20.0,>=0.8->apache-beam==2.30.0.dev0) (2.4.7)
Collecting pbr>=0.11
  Using cached pbr-5.5.1-py2.py3-none-any.whl (106 kB)
Collecting isodate>=0.6.0
  Using cached isodate-0.6.0-py2.py3-none-any.whl (45 kB)
Collecting requests-oauthlib>=0.5.0
  Using cached requests_oauthlib-1.3.0-py2.py3-none-any.whl (23 kB)
Collecting certifi>=2017.4.17
  Using cached certifi-2020.12.5-py2.py3-none-any.whl (147 kB)
Collecting pyasn1>=0.1.7
  Using cached pyasn1-0.4.8-py2.py3-none-any.whl (77 kB)
Collecting more-itertools>=4.0.0
  Using cached more_itertools-8.7.0-py3-none-any.whl (48 kB)
Requirement already satisfied: py>=1.5.0 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from pytest<5.0,>=4.4.0->apache-beam==2.30.0.dev0) (1.10.0)
Requirement already satisfied: pluggy<1.0,>=0.12 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from pytest<5.0,>=4.4.0->apache-beam==2.30.0.dev0) (0.13.1)
Collecting attrs>=17.4.0
  Using cached attrs-20.3.0-py2.py3-none-any.whl (49 kB)
Collecting wcwidth
  Using cached wcwidth-0.2.5-py2.py3-none-any.whl (30 kB)
Collecting atomicwrites>=1.0
  Using cached atomicwrites-1.4.0-py2.py3-none-any.whl (6.8 kB)
Requirement already satisfied: importlib-metadata>=0.12 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from pytest<5.0,>=4.4.0->apache-beam==2.30.0.dev0) (2.1.1)
Requirement already satisfied: zipp>=0.5 in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.6/site-packages> (from importlib-metadata>=0.12->pytest<5.0,>=4.4.0->apache-beam==2.30.0.dev0) (3.4.1)
Collecting execnet>=1.1
  Using cached execnet-1.8.0-py2.py3-none-any.whl (39 kB)
Collecting pytest-forked
  Using cached pytest_forked-1.3.0-py2.py3-none-any.whl (4.7 kB)
Collecting apipkg>=1.4
  Using cached apipkg-1.5-py2.py3-none-any.whl (4.9 kB)
Collecting chardet<5,>=3.0.2
  Using cached chardet-4.0.0-py2.py3-none-any.whl (178 kB)
Collecting idna<3,>=2.5
  Using cached idna-2.10-py2.py3-none-any.whl (58 kB)
Collecting oauthlib>=3.0.0
  Using cached oauthlib-3.1.0-py2.py3-none-any.whl (147 kB)
Collecting greenlet!=0.4.17
  Using cached greenlet-1.0.0-cp36-cp36m-manylinux2010_x86_64.whl (156 kB)
Collecting wrapt
  Using cached wrapt-1.12.1-cp36-cp36m-linux_x86_64.whl
Collecting docker
  Using cached docker-5.0.0-py2.py3-none-any.whl (146 kB)
Collecting deprecation
  Using cached deprecation-2.1.0-py2.py3-none-any.whl (11 kB)
Collecting websocket-client>=0.32.0
  Using cached websocket_client-0.58.0-py2.py3-none-any.whl (61 kB)
Building wheels for collected packages: apache-beam
  Building wheel for apache-beam (setup.py): started
  Building wheel for apache-beam (setup.py): finished with status 'done'
  Created wheel for apache-beam: filename=apache_beam-2.30.0.dev0-py3-none-any.whl size=2453766 sha256=dda345ac9e0b9ac97a775b391c58122bbff50db07274b96d0c13e031490cf635
  Stored in directory: /home/jenkins/.cache/pip/wheels/b5/87/49/3d27e2cd668283bc665894d7d115f547701ab263ea978b116e
Successfully built apache-beam
Installing collected packages: pyasn1, urllib3, rsa, pycparser, pyasn1-modules, idna, chardet, certifi, cachetools, wcwidth, requests, pytz, python-dateutil, oauthlib, more-itertools, jmespath, googleapis-common-protos, google-auth, cffi, attrs, atomicwrites, websocket-client, requests-oauthlib, pytest, numpy, isodate, httplib2, grpcio-gcp, google-crc32c, google-api-core, docopt, botocore, apipkg, wrapt, typing-extensions, s3transfer, pytest-forked, pymongo, pydot, pyarrow, proto-plus, pbr, oauth2client, nose, msrest, hdfs, grpc-google-iam-v1, greenlet, google-resumable-media, google-cloud-core, fasteners, fastavro, execnet, docker, dill, deprecation, dataclasses, cryptography, crcmod, azure-core, avro-python3, testcontainers, tenacity, sqlalchemy, requests-mock, pyyaml, pytest-xdist, pytest-timeout, pyhamcrest, psycopg2-binary, parameterized, pandas, nose-xunitmp, mock, google-cloud-vision, google-cloud-videointelligence, google-cloud-spanner, google-cloud-pubsub, google-cloud-language, google-cloud-dlp, google-cloud-datastore, google-cloud-bigtable, google-cloud-bigquery, google-apitools, freezegun, boto3, azure-storage-blob, apache-beam

> Task :release:go-licenses:py:dockerRun
# cd .; git clone -- https://go.googlesource.com/text /go/src/golang.org/x/text
Cloning into '/go/src/golang.org/x/text'...
error: RPC failed; curl 56 GnuTLS recv error (-54): Error in the pull function.
fatal: the remote end hung up unexpectedly
fatal: early EOF
fatal: index-pack failed
package golang.org/x/text/secure/bidirule: exit status 128
cannot find package "golang.org/x/text/unicode/bidi" in any of:
	/usr/local/go/src/golang.org/x/text/unicode/bidi (from $GOROOT)
	/go/src/golang.org/x/text/unicode/bidi (from $GOPATH)
cannot find package "golang.org/x/text/unicode/norm" in any of:
	/usr/local/go/src/golang.org/x/text/unicode/norm (from $GOROOT)
	/go/src/golang.org/x/text/unicode/norm (from $GOPATH)

> Task :release:go-licenses:java:dockerRun
+ go-licenses save github.com/apache/beam/sdks/java/container --save_path=/output/licenses

> Task :sdks:go:resolveBuildDependencies
Resolving golang.org/x/text: commit='23ae387dee1f90d29a23c0e87ee0b46038fbed0e', urls=[https://go.googlesource.com/text]
Resolving cached github.com/etcd-io/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/etcd-io/etcd.git, git@github.com:etcd-io/etcd.git]

> Task :release:go-licenses:java:dockerRun
+ go-licenses csv github.com/apache/beam/sdks/java/container
+ tee /output/licenses/list.csv
github.com/golang/protobuf,https://github.com/golang/protobuf/blob/master/LICENSE,BSD-3-Clause
google.golang.org/protobuf,https://go.googlesource.com/protobuf/+/refs/heads/master/LICENSE,BSD-3-Clause
google.golang.org/grpc,https://github.com/grpc/grpc-go/blob/master/LICENSE,Apache-2.0
google.golang.org/genproto/googleapis/rpc/status,https://github.com/googleapis/go-genproto/blob/master/LICENSE,Apache-2.0
golang.org/x/text,https://go.googlesource.com/text/+/refs/heads/master/LICENSE,BSD-3-Clause
github.com/apache/beam/sdks/go/pkg/beam,https://github.com/apache/beam/blob/master/sdks/go/README.md,Apache-2.0
golang.org/x/net,https://go.googlesource.com/net/+/refs/heads/master/LICENSE,BSD-3-Clause
golang.org/x/sys,https://go.googlesource.com/sys/+/refs/heads/master/LICENSE,BSD-3-Clause
github.com/apache/beam/sdks/java/container,https://github.com/apache/beam/blob/master/LICENSE,Apache-2.0
+ chmod -R a+w /output/licenses

> Task :release:go-licenses:py:dockerRun FAILED
> Task :release:go-licenses:java:createLicenses
> Task :sdks:java:container:java8:copyGolangLicenses

> Task :sdks:go:resolveBuildDependencies
Resolving google.golang.org/api: commit='0324d5e90dc7753607860272666845fad9ceb97e', urls=[https://code.googlesource.com/google-api-go-client]
Resolving google.golang.org/genproto: commit='4d944d34d83c502a5f761500a14d8842648415c3', urls=[https://github.com/google/go-genproto]
Resolving google.golang.org/grpc: commit='5e8f83304c0563d1ba74db05fee83d9c18ab9a58', urls=[https://github.com/grpc/grpc-go]
Resolving google.golang.org/protobuf: commit='d165be301fb1e13390ad453281ded24385fd8ebc', urls=[https://go.googlesource.com/protobuf]
Resolving cached github.com/etcd-io/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/etcd-io/etcd.git, git@github.com:etcd-io/etcd.git]
Resolving cached github.com/etcd-io/etcd: commit='11214aa33bf5a47d3d9d8dafe0f6b97237dfe921', urls=[https://github.com/etcd-io/etcd.git, git@github.com:etcd-io/etcd.git]

> Task :sdks:go:installDependencies
> Task :sdks:go:buildLinuxAmd64
> Task :sdks:go:goBuild

> Task :sdks:java:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/go>

> Task :sdks:python:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/go>

> Task :sdks:java:container:installDependencies
> Task :sdks:python:container:installDependencies

> Task :sdks:java:container:buildLinuxAmd64
Unable to watch the file system for changes. The inotify watches limit is too low.

> Task :sdks:java:container:goBuild
> Task :sdks:python:container:buildDarwinAmd64
> Task :sdks:java:container:java8:copySdkHarnessLauncher
> Task :sdks:python:container:buildLinuxAmd64
> Task :sdks:python:container:goBuild
> Task :sdks:python:container:py36:copyLauncherDependencies
> Task :sdks:java:container:java8:dockerPrepare
> Task :sdks:java:container:java8:docker
> Task :runners:spark:2:job-server:sparkJobServerCleanup

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':release:go-licenses:py:dockerRun'.
> Process 'command 'docker'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 11m 56s
171 actionable tasks: 120 executed, 47 from cache, 4 up-to-date
Gradle was unable to watch the file system for changes. The inotify watches limit is too low.

Publishing build scan...
https://gradle.com/s/pkqbvafk6y7e6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark #2207

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/2207/display/redirect?page=changes>

Changes:

[sychen] Add a BQ option for configuring buffering duration when auto-sharding is

[suztomo] [BEAM-11010] Upgrading google-cloud-pubsublite to 0.13.2

[suztomo] [BEAM-11010] Copying SubscriberOptions from pubsublite repo

[suztomo] [BEAM-11010] Declaring flogger-system-backend to avoid conflicts

[Kyle Weaver] [BEAM-12194] Enable SqlTransform::registerUdaf in ZetaSQL.

[Kyle Weaver] [BEAM-12194] Code style changes from review.

[Boyuan Zhang] [BEAM-12114] Dataflow should apply KAFKA_READ_OVERRIDE when it's not

[noreply] Make sdk/worker_harness_container_image fully backwards compatible


------------------------------------------
[...truncated 989.49 KB...]
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: java.lang.NullPointerException
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_windowing_before_sql (apache_beam.transforms.sql_test.SqlTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/transforms/sql_test.py",> line 179, in test_windowing_before_sql
    assert_that(out, equal_to([(1, ), (1, ), (1, )]))
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 582, in __exit__
    self.result = self.run()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 112, in run
    state = result.wait_until_finish()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 602, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline BeamApp-jenkins-0422003708-2ad8430e_5d102184-af89-4c8e-beea-346f647b5c54 failed in state FAILED: java.lang.NullPointerException
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar'> '37615']
root: DEBUG: Waiting for grpc channel to be ready at localhost:37615.
apache_beam.utils.subprocess_server: INFO: b'Starting expansion service at localhost:37615'
root: DEBUG: Waiting for grpc channel to be ready at localhost:37615.
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:36:59 AM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
apache_beam.utils.subprocess_server: INFO: b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5680a178'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5fdef03a'
root: DEBUG: Waiting for grpc channel to be ready at localhost:37615.
root: DEBUG: Waiting for grpc channel to be ready at localhost:37615.
root: DEBUG: Waiting for grpc channel to be ready at localhost:37615.
root: DEBUG: Waiting for grpc channel to be ready at localhost:37615.
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:00 AM org.apache.beam.sdk.expansion.service.ExpansionService expand'
apache_beam.utils.subprocess_server: INFO: b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:01 AM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
apache_beam.utils.subprocess_server: INFO: b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:03 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
apache_beam.utils.subprocess_server: INFO: b'INFO: SQL:'
apache_beam.utils.subprocess_server: INFO: b'SELECT COUNT(*) AS `count`'
apache_beam.utils.subprocess_server: INFO: b'FROM `beam`.`PCOLLECTION` AS `PCOLLECTION`'
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:03 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
apache_beam.utils.subprocess_server: INFO: b'INFO: SQLPlan>'
apache_beam.utils.subprocess_server: INFO: b'LogicalAggregate(group=[{}], count=[COUNT()])'
apache_beam.utils.subprocess_server: INFO: b'  LogicalProject($f0=[0])'
apache_beam.utils.subprocess_server: INFO: b'    BeamIOSourceRel(table=[[beam, PCOLLECTION]])'
apache_beam.utils.subprocess_server: INFO: b''
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:03 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
apache_beam.utils.subprocess_server: INFO: b'INFO: BEAMPlan>'
apache_beam.utils.subprocess_server: INFO: b'BeamAggregationRel(group=[{}], count=[COUNT()])'
apache_beam.utils.subprocess_server: INFO: b'  BeamIOSourceRel(table=[[beam, PCOLLECTION]])'
apache_beam.utils.subprocess_server: INFO: b''
root: DEBUG: Sending SIGINT to job_server
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: No image given, using default Python SDK image
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: Python SDK container image set to "apache/beam_python3.6_sdk:2.30.0.dev" for Docker environment
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f0a2e1f0ea0> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 27 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_Create-Impulse_3\n  Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-FlatMap-lambda-at-core-py-2955-_4\n  Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-AddRandomKeys_7\n  Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-Map-reify_timestamps-_9\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-GroupByKey_10\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-FlatMap-restore_timestamps-_11\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-RemoveRandomKeys_12\n  Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-Map-decode-_13\n  Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map-lambda-at-sql_test-py-174-_14\n  Map(<lambda at sql_test.py:174>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_WindowInto-WindowIntoFn-_15\n  WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-sele\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/selectKeys/AddKeys/Map/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-Grou\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-Combine-Pa\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/Combine/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToRow-ParM\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToRow/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-mergeRecord-ParMultiDo-Anonymous-\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/mergeRecord/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_19\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_20\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_22\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_23\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_24\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_26\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_27\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_28\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_29\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_30\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_31\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_32\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function sort_stages at 0x7f0a2e1f1620> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 27 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_Create-Impulse_3\n  Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-FlatMap-lambda-at-core-py-2955-_4\n  Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-AddRandomKeys_7\n  Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-Map-reify_timestamps-_9\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-GroupByKey_10\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-ReshufflePerKey-FlatMap-restore_timestamps-_11\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-MaybeReshuffle-Reshuffle-RemoveRandomKeys_12\n  Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create-Map-decode-_13\n  Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map-lambda-at-sql_test-py-174-_14\n  Map(<lambda at sql_test.py:174>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_WindowInto-WindowIntoFn-_15\n  WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-sele\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/selectKeys/AddKeys/Map/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToKvs-Grou\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-Combine-Pa\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/Combine/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-Group-CombineFieldsByFields-ToRow-ParM\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToRow/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform-beam-external-java-sql-v1--BeamAggregationRel_40-mergeRecord-ParMultiDo-Anonymous-\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/mergeRecord/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_19\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_20\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_22\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_23\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_24\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_26\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_27\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_28\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_29\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_30\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_31\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_32\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'experiments' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_name' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'runner' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'temp_location' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'dataflow_kms_key' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'enable_streaming_engine' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'project' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_region' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'streaming' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_cache_millis' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_options' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_endpoint' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'output_executable_path' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'sdk_worker_parallelism' was already added
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: java.lang.NullPointerException
	at org.apache.beam.runners.spark.SparkPipelineRunner.run(SparkPipelineRunner.java:120)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: java.lang.NullPointerException
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_zetasql_generate_data (apache_beam.transforms.sql_test.SqlTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/transforms/sql_test.py",> line 164, in test_zetasql_generate_data
    assert_that(out, equal_to([(1, "foo", 3.14)]))
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/pipeline.py",> line 582, in __exit__
    self.result = self.run()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 112, in run
    state = result.wait_until_finish()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 602, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline BeamApp-jenkins-0422003718-e19450c_05eaba02-eb2c-4255-ad16-ea50704ca3e0 failed in state FAILED: java.lang.NullPointerException
-------------------- >> begin captured logging << --------------------
apache_beam.utils.subprocess_server: INFO: Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar>
apache_beam.utils.subprocess_server: INFO: Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.30.0-SNAPSHOT.jar'> '34521']
root: DEBUG: Waiting for grpc channel to be ready at localhost:34521.
apache_beam.utils.subprocess_server: INFO: b'Starting expansion service at localhost:34521'
root: DEBUG: Waiting for grpc channel to be ready at localhost:34521.
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:08 AM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
apache_beam.utils.subprocess_server: INFO: b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5680a178'
apache_beam.utils.subprocess_server: INFO: b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$3/1130478920@5fdef03a'
root: DEBUG: Waiting for grpc channel to be ready at localhost:34521.
root: DEBUG: Waiting for grpc channel to be ready at localhost:34521.
root: DEBUG: Waiting for grpc channel to be ready at localhost:34521.
root: DEBUG: Waiting for grpc channel to be ready at localhost:34521.
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:09 AM org.apache.beam.sdk.expansion.service.ExpansionService expand'
apache_beam.utils.subprocess_server: INFO: b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:10 AM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
apache_beam.utils.subprocess_server: INFO: b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
apache_beam.utils.subprocess_server: INFO: b'Apr 22, 2021 12:37:14 AM org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner convertToBeamRelInternal'
apache_beam.utils.subprocess_server: INFO: b'INFO: BEAMPlan>'
apache_beam.utils.subprocess_server: INFO: b"BeamZetaSqlCalcRel(expr#0=[{inputs}], expr#1=[1:BIGINT], expr#2=['foo':VARCHAR], expr#3=[3.1400000000000001243E0:DOUBLE], int=[$t1], str=[$t2], flt=[$t3])"
apache_beam.utils.subprocess_server: INFO: b'  BeamValuesRel(tuples=[[{ 0 }]])'
apache_beam.utils.subprocess_server: INFO: b''
root: DEBUG: Sending SIGINT to job_server
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: No image given, using default Python SDK image
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.6_sdk:2.30.0.dev
root: INFO: Python SDK container image set to "apache/beam_python3.6_sdk:2.30.0.dev" for Docker environment
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function lift_combiners at 0x7f0a2e1f0ea0> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 16 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--Impulse\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Outp\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Boun\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamZetaSqlCalcRel_17-ParDo-Calc--ParMultiDo-Calc-\n  SqlTransform(beam:external:java:sql:v1)/BeamZetaSqlCalcRel_17/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_5\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_6\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_8\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_9\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_10\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_12\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_13\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_14\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_15\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_16\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_17\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_18\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function sort_stages at 0x7f0a2e1f1620> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 16 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--Impulse\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Outp\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamValuesRel_13-Create-Values-Read-CreateSource--ParDo-Boun\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform-beam-external-java-sql-v1--BeamZetaSqlCalcRel_17-ParDo-Calc--ParMultiDo-Calc-\n  SqlTransform(beam:external:java:sql:v1)/BeamZetaSqlCalcRel_17/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Impulse_5\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-FlatMap-lambda-at-core-py-2955-_6\n  assert_that/Create/FlatMap(<lambda at core.py:2955>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Create-Map-decode-_8\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-WindowInto-WindowIntoFn-_9\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-ToVoidKey_10\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_0_12\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-pair_with_1_13\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Flatten_14\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-GroupByKey_15\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Group-Map-_merge_tagged_vals_under_key-_16\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Unkey_17\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that-Match_18\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'experiments' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_name' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'runner' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'temp_location' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'dataflow_kms_key' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'enable_streaming_engine' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'project' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_region' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'worker_zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'zone' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'streaming' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_cache_millis' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'environment_options' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'job_endpoint' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'output_executable_path' was already added
apache_beam.runners.portability.portable_runner: DEBUG: Runner option 'sdk_worker_parallelism' was already added
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to RUNNING
root: DEBUG: java.lang.NullPointerException
	at org.apache.beam.runners.spark.SparkPipelineRunner.run(SparkPipelineRunner.java:120)
	at org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

root: ERROR: java.lang.NullPointerException
apache_beam.runners.portability.portable_runner: INFO: Job state changed to FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-xlangSqlValidateRunner.xml
----------------------------------------------------------------------
XML: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 9 tests in 105.030s

FAILED (errors=9)

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:2:job-server:sparkJobServerCleanup

FAILURE: Build completed with 3 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 36m 55s
182 actionable tasks: 136 executed, 42 from cache, 4 up-to-date
Gradle was unable to watch the file system for changes. The inotify watches limit is too low.

Publishing build scan...
https://gradle.com/s/rrlwfo4ha4dxe

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org