You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/10/21 18:51:56 UTC

Build failed in Jenkins: beam_PostCommit_XVR_Flink #3157

See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/3157/display/redirect?page=changes>

Changes:

[neville.lyh] [BEAM-10787] make external sorter 2048MB memory limit Hadoop only

[sjvanrossum] Change imports of org.codehaus.jackson to com.fasterxml.jackson


------------------------------------------
[...truncated 488.41 KB...]
Building dependency tree...
Reading state information...
libyaml-dev is already the newest version (0.2.1-1).
Suggested packages:
  distcc | icecc
The following NEW packages will be installed:
  ccache libsnappy-dev libsnappy1v5
0 upgraded, 3 newly installed, 0 to remove and 5 not upgraded.
Need to get 203 kB of archives.
After this operation, 514 kB of additional disk space will be used.
Get:1 http://deb.debian.org/debian buster/main amd64 ccache amd64 3.6-1 [156 kB]
Get:2 http://deb.debian.org/debian buster/main amd64 libsnappy1v5 amd64 1.1.7-1 [17.0 kB]
Get:3 http://deb.debian.org/debian buster/main amd64 libsnappy-dev amd64 1.1.7-1 [29.2 kB]
debconf: delaying package configuration, since apt-utils is not installed
Fetched 203 kB in 0s (2638 kB/s)
Selecting previously unselected package ccache.
(Reading database ... 
(Reading database ... 5%(Reading database ... 10%(Reading database ... 15%(Reading database ... 20%(Reading database ... 25%(Reading database ... 30%(Reading database ... 35%(Reading database ... 40%(Reading database ... 45%(Reading database ... 50%(Reading database ... 55%(Reading database ... 60%(Reading database ... 65%(Reading database ... 70%
(Reading database ... 75%
(Reading database ... 80%
(Reading database ... 85%
(Reading database ... 90%
(Reading database ... 95%
(Reading database ... 100%(Reading database ... 24603 files and directories currently installed.)
Preparing to unpack .../ccache_3.6-1_amd64.deb ...
Unpacking ccache (3.6-1) ...
Selecting previously unselected package libsnappy1v5:amd64.
Preparing to unpack .../libsnappy1v5_1.1.7-1_amd64.deb ...
Unpacking libsnappy1v5:amd64 (1.1.7-1) ...
Selecting previously unselected package libsnappy-dev:amd64.
Preparing to unpack .../libsnappy-dev_1.1.7-1_amd64.deb ...
Unpacking libsnappy-dev:amd64 (1.1.7-1) ...
Setting up ccache (3.6-1) ...
Updating symlinks in /usr/lib/ccache ...
Setting up libsnappy1v5:amd64 (1.1.7-1) ...
Setting up libsnappy-dev:amd64 (1.1.7-1) ...
Processing triggers for libc-bin (2.28-10) ...
Removing intermediate container e72e8f86e0f6
 ---> 29de21ce23d5
Step 6/23 : COPY target/base_image_requirements.txt /tmp/base_image_requirements.txt
 ---> 9cb4c6e51f40
Step 7/23 : RUN     pip install -r /tmp/base_image_requirements.txt &&     python -c "from google.protobuf.internal import api_implementation; assert api_implementation._default_implementation_type == 'cpp'; print ('Verified fast protobuf used.')" &&     rm -rf /root/.cache/pip &&     rm -rf /tmp/base_image_requirements.txt
 ---> Running in fcf9d5d82c47
Ignoring avro: markers 'python_version <= "2.7"' don't match your environment
Ignoring futures: markers 'python_version < "3.0"' don't match your environment
Ignoring pyvcf: markers 'python_version < "3.0"' don't match your environment
Ignoring typing: markers 'python_full_version < "3.5.3"' don't match your environment
Ignoring guppy: markers 'python_version <= "2.7"' don't match your environment
Ignoring numpy: markers 'python_version <= "2.7"' don't match your environment
Ignoring scipy: markers 'python_version <= "2.7"' don't match your environment
Ignoring pandas: markers 'python_version <= "2.7"' don't match your environment
Ignoring tensorflow: markers 'python_version <= "2.7"' don't match your environment
Collecting avro-python3==1.8.2
  Downloading avro-python3-1.8.2.tar.gz (36 kB)
Collecting fastavro==1.0.0.post1
  Downloading fastavro-1.0.0.post1-cp38-cp38-manylinux2014_x86_64.whl (1.7 MB)
Collecting crcmod==1.7
  Downloading crcmod-1.7.tar.gz (89 kB)
Collecting dill==0.3.1.1
  Downloading dill-0.3.1.1.tar.gz (151 kB)
Collecting future==0.18.2
  Downloading future-0.18.2.tar.gz (829 kB)
Collecting grpcio==1.29.0
  Downloading grpcio-1.29.0-cp38-cp38-manylinux2010_x86_64.whl (3.0 MB)
Collecting hdfs==2.5.8
  Downloading hdfs-2.5.8.tar.gz (41 kB)
Collecting httplib2==0.12.0
  Downloading httplib2-0.12.0.tar.gz (218 kB)
Collecting mock==2.0.0
  Downloading mock-2.0.0-py2.py3-none-any.whl (56 kB)
Collecting oauth2client==4.1.3
  Downloading oauth2client-4.1.3-py2.py3-none-any.whl (98 kB)
Collecting protobuf==3.12.2
  Downloading protobuf-3.12.2-cp38-cp38-manylinux1_x86_64.whl (1.3 MB)
Collecting pyarrow==0.16.0
  Downloading pyarrow-0.16.0-cp38-cp38-manylinux2014_x86_64.whl (63.2 MB)
Collecting pydot==1.4.1
  Downloading pydot-1.4.1-py2.py3-none-any.whl (19 kB)
Collecting pytz==2019.3
  Downloading pytz-2019.3-py2.py3-none-any.whl (509 kB)
Collecting pyyaml==5.1
  Downloading PyYAML-5.1.tar.gz (274 kB)
Collecting typing-extensions==3.7.4.3
  Downloading typing_extensions-3.7.4.3-py3-none-any.whl (22 kB)
Collecting google-api-core==1.22.0
  Downloading google_api_core-1.22.0-py2.py3-none-any.whl (91 kB)

> Task :sdks:java:container:pullLicenses

> Task :sdks:java:container:generateLicenseReport

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 40s
1 actionable task: 1 executed

Publishing build scan...
https://gradle.com/s/gepdigim7posu

Already using interpreter /usr/bin/python3
Using base prefix '/usr'
New python executable in <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/build/virtualenv/bin/python3>
Not overwriting existing python script <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/build/virtualenv/bin/python> (you must use <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/build/virtualenv/bin/python3)>
Installing setuptools, pip, wheel...
done.
DEPRECATION: Python 3.5 reached the end of its life on September 13th, 2020. Please upgrade your Python as Python 3.5 is no longer maintained. pip 21.0 will drop support for Python 3.5 in January 2021. pip 21.0 will remove support for this functionality.
Requirement already satisfied: beautifulsoup4<5.0,>=4.9.0 in ./sdks/java/container/build/virtualenv/lib/python3.5/site-packages (from -r <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/requirement.txt> (line 19)) (4.9.3)
Requirement already satisfied: future<1.0.0,>=0.16.0 in ./sdks/java/container/build/virtualenv/lib/python3.5/site-packages (from -r <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/requirement.txt> (line 20)) (0.18.2)
Requirement already satisfied: pyyaml<6.0.0,>=3.12 in ./sdks/java/container/build/virtualenv/lib/python3.5/site-packages (from -r <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/requirement.txt> (line 21)) (5.3.1)
Requirement already satisfied: tenacity<6.0,>=5.0.2 in ./sdks/java/container/build/virtualenv/lib/python3.5/site-packages (from -r <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/requirement.txt> (line 22)) (5.1.5)
Requirement already satisfied: soupsieve>1.2; python_version >= "3.0" in ./sdks/java/container/build/virtualenv/lib/python3.5/site-packages (from beautifulsoup4<5.0,>=4.9.0->-r <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/requirement.txt> (line 19)) (2.0.1)
Requirement already satisfied: six>=1.9.0 in ./sdks/java/container/build/virtualenv/lib/python3.5/site-packages (from tenacity<6.0,>=5.0.2->-r <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/requirement.txt> (line 22)) (1.15.0)
Executing <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/build/virtualenv/bin/python> <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py> --license_dir=<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/java_third_party_licenses>        --dep_url_yaml=<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> 
INFO:root:Pulling license for 180 dependencies using 16 threads.

> Task :sdks:python:container:py38:docker FAILED
Collecting google-apitools==0.5.28
  Downloading google-apitools-0.5.28.tar.gz (172 kB)
Collecting google-cloud-pubsub==1.0.2
  Downloading google_cloud_pubsub-1.0.2-py2.py3-none-any.whl (118 kB)
Collecting google-cloud-bigquery==1.26.1
  Downloading google_cloud_bigquery-1.26.1-py2.py3-none-any.whl (170 kB)
Collecting google-cloud-bigtable==1.0.0
  Downloading google_cloud_bigtable-1.0.0-py2.py3-none-any.whl (232 kB)
Collecting google-cloud-core==1.4.1
  Downloading google_cloud_core-1.4.1-py2.py3-none-any.whl (26 kB)
Collecting google-cloud-datastore==1.7.4
  Downloading google_cloud_datastore-1.7.4-py2.py3-none-any.whl (82 kB)
Collecting cython==0.29.13
  Downloading Cython-0.29.13.tar.gz (2.1 MB)
Collecting guppy3==3.0.9
  Downloading guppy3-3.0.9.tar.gz (333 kB)
Collecting mmh3==2.5.1
  Downloading mmh3-2.5.1.tar.gz (9.8 kB)
Collecting numpy==1.17.3
  Downloading numpy-1.17.3-cp38-cp38-manylinux1_x86_64.whl (20.5 MB)
Collecting scipy==1.4.1
  Downloading scipy-1.4.1-cp38-cp38-manylinux1_x86_64.whl (26.0 MB)
Collecting pandas==0.25.2
  Downloading pandas-0.25.2-cp38-cp38-manylinux1_x86_64.whl (10.4 MB)
Collecting protorpc==0.12.0
  Downloading protorpc-0.12.0.tar.gz (117 kB)
Collecting python-gflags==3.0.6
  Downloading python-gflags-3.0.6.tar.gz (49 kB)
Collecting tensorflow==2.3.0
  Downloading tensorflow-2.3.0-cp38-cp38-manylinux2010_x86_64.whl (320.5 MB)
ERROR: Exception:
Traceback (most recent call last):
  File "/usr/local/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py", line 437, in _error_catcher
    yield
  File "/usr/local/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py", line 519, in read
    data = self._fp.read(amt) if not fp_closed else b""
  File "/usr/local/lib/python3.8/site-packages/pip/_vendor/cachecontrol/filewrapper.py", line 62, in read
    data = self.__fp.read(amt)
  File "/usr/local/lib/python3.8/http/client.py", line 458, in read
    n = self.readinto(b)
  File "/usr/local/lib/python3.8/http/client.py", line 502, in readinto
    n = self.fp.readinto(b)
  File "/usr/local/lib/python3.8/socket.py", line 669, in readinto
    return self._sock.recv_into(b)
  File "/usr/local/lib/python3.8/ssl.py", line 1241, in recv_into
    return self.read(nbytes, buffer)
  File "/usr/local/lib/python3.8/ssl.py", line 1099, in read
    return self._sslobj.read(len, buffer)
ConnectionResetError: [Errno 104] Connection reset by peer

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/cli/base_command.py", line 228, in _main
    status = self.run(options, args)
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/cli/req_command.py", line 182, in wrapper
    return func(self, options, args)
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/commands/install.py", line 323, in run
    requirement_set = resolver.resolve(
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/resolution/legacy/resolver.py", line 183, in resolve
    discovered_reqs.extend(self._resolve_one(requirement_set, req))
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/resolution/legacy/resolver.py", line 388, in _resolve_one
    abstract_dist = self._get_abstract_dist_for(req_to_install)
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/resolution/legacy/resolver.py", line 340, in _get_abstract_dist_for
    abstract_dist = self.preparer.prepare_linked_requirement(req)
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/operations/prepare.py", line 467, in prepare_linked_requirement
    local_file = unpack_url(
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/operations/prepare.py", line 255, in unpack_url
    file = get_http_url(
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/operations/prepare.py", line 129, in get_http_url
    from_path, content_type = _download_http_url(
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/operations/prepare.py", line 282, in _download_http_url
    for chunk in download.chunks:
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/cli/progress_bars.py", line 168, in iter
    for x in it:
  File "/usr/local/lib/python3.8/site-packages/pip/_internal/network/utils.py", line 64, in response_chunks
    for chunk in response.raw.stream(
  File "/usr/local/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py", line 576, in stream
    data = self.read(amt=amt, decode_content=decode_content)
  File "/usr/local/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py", line 541, in read
    raise IncompleteRead(self._fp_bytes_read, self.length_remaining)
  File "/usr/local/lib/python3.8/contextlib.py", line 131, in __exit__
    self.gen.throw(type, value, traceback)
  File "/usr/local/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py", line 455, in _error_catcher
    raise ProtocolError("Connection broken: %r" % e, e)
pip._vendor.urllib3.exceptions.ProtocolError: ("Connection broken: ConnectionResetError(104, 'Connection reset by peer')", ConnectionResetError(104, 'Connection reset by peer'))
The command '/bin/sh -c pip install -r /tmp/base_image_requirements.txt &&     python -c "from google.protobuf.internal import api_implementation; assert api_implementation._default_implementation_type == 'cpp'; print ('Verified fast protobuf used.')" &&     rm -rf /root/.cache/pip &&     rm -rf /tmp/base_image_requirements.txt' returned a non-zero code: 2

> Task :sdks:java:container:pullLicenses
INFO:root:pull_licenses_java.py succeed. It took 14.288783 seconds with 16 threads.
Copy licenses to <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/container/build/target/third_party_licenses.>
Finished license_scripts.sh

> Task :sdks:java:container:copyDockerfileDependencies UP-TO-DATE

> Task :sdks:java:container:resolveBuildDependencies
Resolving ./github.com/apache/beam/sdks/go@<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/go>

> Task :sdks:java:container:installDependencies
> Task :sdks:java:container:buildLinuxAmd64 UP-TO-DATE
> Task :sdks:java:container:goBuild
> Task :sdks:java:container:dockerPrepare
> Task :sdks:java:container:docker

> Task :runners:flink:1.10:job-server:flinkJobServerCleanup
Stopping job server pid: 4387.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:python:container:py38:docker'.
> Process 'command 'docker'' finished with non-zero exit value 2

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 6s
142 actionable tasks: 28 executed, 114 up-to-date

Publishing build scan...
https://gradle.com/s/oq6q2j3sijcrk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
No JDK named ‘JDK 1.8 (latest)’ found

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_XVR_Flink #3159

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/3159/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Flink #3158

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/3158/display/redirect?page=changes>

Changes:

[Kenneth Knowles] Make streaming and "auto" Dataflow ValidatesRunner jobs have different

[Kenneth Knowles] Upgrade checkerframework to 3.7.0

[Kenneth Knowles] Upgrade checkerframework gradle plugin to 0.5.11

[Boyuan Zhang] Lengthprefix any input coder for an ProcessBundleDescriptor.

[noreply] Merge pull request #13153 from [BEAM-11093] Adding BQ snippets for

[noreply] [BEAM-11094] Only work with fully expanded PCollections in

[noreply] Use a newer Python interpreter version as default. (#13140)


------------------------------------------
[...truncated 318.17 KB...]
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'flink_master' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'pubsub_root_url' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_tagged_join (apache_beam.transforms.sql_test.SqlTransformTest) ... INFO:apache_beam.utils.subprocess_server:Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.26.0-SNAPSHOT.jar>
INFO:apache_beam.utils.subprocess_server:Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.26.0-SNAPSHOT.jar'> '59427']
DEBUG:root:Waiting for grpc channel to be ready at localhost:59427.
INFO:apache_beam.utils.subprocess_server:b'Starting expansion service at localhost:59427'
DEBUG:root:Waiting for grpc channel to be ready at localhost:59427.
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:45:51 AM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
INFO:apache_beam.utils.subprocess_server:b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
INFO:apache_beam.utils.subprocess_server:b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/2121744517@626b2d4a'
INFO:apache_beam.utils.subprocess_server:b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/2121744517@5e91993f'
DEBUG:root:Waiting for grpc channel to be ready at localhost:59427.
DEBUG:root:Waiting for grpc channel to be ready at localhost:59427.
DEBUG:root:Waiting for grpc channel to be ready at localhost:59427.
DEBUG:root:Waiting for grpc channel to be ready at localhost:59427.
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:45:52 AM org.apache.beam.sdk.expansion.service.ExpansionService expand'
INFO:apache_beam.utils.subprocess_server:b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:45:53 AM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
INFO:apache_beam.utils.subprocess_server:b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:45:55 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
INFO:apache_beam.utils.subprocess_server:b'INFO: SQL:'
INFO:apache_beam.utils.subprocess_server:b'SELECT `simple`.`id` AS `id`, `enrich`.`metadata` AS `metadata`'
INFO:apache_beam.utils.subprocess_server:b'FROM `beam`.`simple` AS `simple`'
INFO:apache_beam.utils.subprocess_server:b'INNER JOIN `beam`.`enrich` AS `enrich` ON `simple`.`id` = `enrich`.`id`'
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:45:55 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
INFO:apache_beam.utils.subprocess_server:b'INFO: SQLPlan>'
INFO:apache_beam.utils.subprocess_server:b'LogicalProject(id=[$0], metadata=[$4])'
INFO:apache_beam.utils.subprocess_server:b'  LogicalJoin(condition=[=($0, $3)], joinType=[inner])'
INFO:apache_beam.utils.subprocess_server:b'    BeamIOSourceRel(table=[[beam, simple]])'
INFO:apache_beam.utils.subprocess_server:b'    BeamIOSourceRel(table=[[beam, enrich]])'
INFO:apache_beam.utils.subprocess_server:b''
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:45:55 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
INFO:apache_beam.utils.subprocess_server:b'INFO: BEAMPlan>'
INFO:apache_beam.utils.subprocess_server:b'BeamCalcRel(expr#0..4=[{inputs}], id=[$t2], metadata=[$t1])'
INFO:apache_beam.utils.subprocess_server:b'  BeamCoGBKJoinRel(condition=[=($2, $0)], joinType=[inner])'
INFO:apache_beam.utils.subprocess_server:b'    BeamIOSourceRel(table=[[beam, enrich]])'
INFO:apache_beam.utils.subprocess_server:b'    BeamIOSourceRel(table=[[beam, simple]])'
INFO:apache_beam.utils.subprocess_server:b''
DEBUG:root:Sending SIGINT to job_server
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.6_sdk:2.26.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.6_sdk:2.26.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f95dbda8bf8> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:42 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Create enrich/Impulse_3\n  Create enrich/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/FlatMap(<lambda at core.py:2956>)_4\n  Create enrich/FlatMap(<lambda at core.py:2956>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n  Create enrich/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n  Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n  Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_11\n  Create enrich/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/MaybeReshuffle/Reshuffle/RemoveRandomKeys_12\n  Create enrich/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create enrich/Map(decode)_13\n  Create enrich/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/Impulse_15\n  Create simple/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/FlatMap(<lambda at core.py:2956>)_16\n  Create simple/FlatMap(<lambda at core.py:2956>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/AddRandomKeys_19\n  Create simple/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_21\n  Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_22\n  Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_23\n  Create simple/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/MaybeReshuffle/Reshuffle/RemoveRandomKeys_24\n  Create simple/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create simple/Map(decode)_25\n  Create simple/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_6/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_6/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_5/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_5/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/left_TimestampCombiner/Flatten.PCollections\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/left_TimestampCombiner/Flatten.PCollections:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/right_TimestampCombiner/Flatten.PCollections\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/right_TimestampCombiner/Flatten.PCollections:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeylhs/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeylhs/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeyrhs/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/extractKeyrhs/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/Flatten\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/GBK\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/GBK:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/ParDo(ConvertCoGbkResult)/ParMultiDo(ConvertCoGbkResult)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Join.Impl/CoGroup.ExpandCrossProduct/ParDo(ConvertCoGbkResult)/ParMultiDo(ConvertCoGbkResult):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Select.Fields/ParDo(Select)/ParMultiDo(Select)\n  SqlTransform(beam:external:java:sql:v1)/BeamCoGBKJoinRel_95/Select.Fields/ParDo(Select)/ParMultiDo(Select):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_7SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_96/ParDo(Calc)/ParMultiDo(Calc)\n  SqlTransform(beam:external:java:sql:v1)/BeamCalcRel_96/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_29\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2956>)_30\n  assert_that/Create/FlatMap(<lambda at core.py:2956>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_32\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_33\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_34\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_36\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_37\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_38\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_39\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_40\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_41\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_42\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'flink_master' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'pubsub_root_url' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_windowing_before_sql (apache_beam.transforms.sql_test.SqlTransformTest) ... INFO:apache_beam.utils.subprocess_server:Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.26.0-SNAPSHOT.jar>
INFO:apache_beam.utils.subprocess_server:Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.26.0-SNAPSHOT.jar'> '58921']
DEBUG:root:Waiting for grpc channel to be ready at localhost:58921.
INFO:apache_beam.utils.subprocess_server:b'Starting expansion service at localhost:58921'
DEBUG:root:Waiting for grpc channel to be ready at localhost:58921.
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:17 AM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
INFO:apache_beam.utils.subprocess_server:b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
INFO:apache_beam.utils.subprocess_server:b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/2121744517@626b2d4a'
INFO:apache_beam.utils.subprocess_server:b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/2121744517@5e91993f'
DEBUG:root:Waiting for grpc channel to be ready at localhost:58921.
DEBUG:root:Waiting for grpc channel to be ready at localhost:58921.
DEBUG:root:Waiting for grpc channel to be ready at localhost:58921.
DEBUG:root:Waiting for grpc channel to be ready at localhost:58921.
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:18 AM org.apache.beam.sdk.expansion.service.ExpansionService expand'
INFO:apache_beam.utils.subprocess_server:b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:19 AM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
INFO:apache_beam.utils.subprocess_server:b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:20 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
INFO:apache_beam.utils.subprocess_server:b'INFO: SQL:'
INFO:apache_beam.utils.subprocess_server:b'SELECT COUNT(*) AS `count`'
INFO:apache_beam.utils.subprocess_server:b'FROM `beam`.`PCOLLECTION` AS `PCOLLECTION`'
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:20 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
INFO:apache_beam.utils.subprocess_server:b'INFO: SQLPlan>'
INFO:apache_beam.utils.subprocess_server:b'LogicalAggregate(group=[{}], count=[COUNT()])'
INFO:apache_beam.utils.subprocess_server:b'  LogicalProject($f0=[0])'
INFO:apache_beam.utils.subprocess_server:b'    BeamIOSourceRel(table=[[beam, PCOLLECTION]])'
INFO:apache_beam.utils.subprocess_server:b''
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:21 AM org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner convertToBeamRel'
INFO:apache_beam.utils.subprocess_server:b'INFO: BEAMPlan>'
INFO:apache_beam.utils.subprocess_server:b'BeamAggregationRel(group=[{}], count=[COUNT()])'
INFO:apache_beam.utils.subprocess_server:b'  BeamIOSourceRel(table=[[beam, PCOLLECTION]])'
INFO:apache_beam.utils.subprocess_server:b''
DEBUG:root:Sending SIGINT to job_server
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.6_sdk:2.26.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.6_sdk:2.26.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f95dbda8bf8> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:29 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Create/Impulse_3\n  Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2956>)_4\n  Create/FlatMap(<lambda at core.py:2956>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n  Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_11\n  Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys_12\n  Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/Map(decode)_13\n  Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at sql_test.py:172>)_14\n  Map(<lambda at sql_test.py:172>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_WindowInto(WindowIntoFn)_15\n  WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_3/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamIOSourceRel_3/Convert.ConvertTransform/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/toRow/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/toRow/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/selectKeys/AddKeys/Map/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/selectKeys/AddKeys/Map/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/GroupByKey\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToKvs/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/Combine/ParDo(Anonymous)/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/Combine/ParDo(Anonymous)/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToRow/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/Group.CombineFieldsByFields/ToRow/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_8SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/mergeRecord/ParMultiDo(Anonymous)\n  SqlTransform(beam:external:java:sql:v1)/BeamAggregationRel_40/mergeRecord/ParMultiDo(Anonymous):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_19\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2956>)_20\n  assert_that/Create/FlatMap(<lambda at core.py:2956>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_22\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_23\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_24\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_26\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_27\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_28\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_29\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_30\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_31\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_32\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'flink_master' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'pubsub_root_url' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_zetasql_generate_data (apache_beam.transforms.sql_test.SqlTransformTest) ... INFO:apache_beam.utils.subprocess_server:Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.26.0-SNAPSHOT.jar>
INFO:apache_beam.utils.subprocess_server:Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.26.0-SNAPSHOT.jar'> '46459']
DEBUG:root:Waiting for grpc channel to be ready at localhost:46459.
INFO:apache_beam.utils.subprocess_server:b'Starting expansion service at localhost:46459'
DEBUG:root:Waiting for grpc channel to be ready at localhost:46459.
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:42 AM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
INFO:apache_beam.utils.subprocess_server:b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
INFO:apache_beam.utils.subprocess_server:b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/2121744517@626b2d4a'
INFO:apache_beam.utils.subprocess_server:b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$1/2121744517@5e91993f'
DEBUG:root:Waiting for grpc channel to be ready at localhost:46459.
DEBUG:root:Waiting for grpc channel to be ready at localhost:46459.
DEBUG:root:Waiting for grpc channel to be ready at localhost:46459.
DEBUG:root:Waiting for grpc channel to be ready at localhost:46459.
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:43 AM org.apache.beam.sdk.expansion.service.ExpansionService expand'
INFO:apache_beam.utils.subprocess_server:b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
INFO:apache_beam.utils.subprocess_server:b'Oct 22, 2020 12:46:44 AM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
INFO:apache_beam.utils.subprocess_server:b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
DEBUG:root:Sending SIGINT to job_server
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
DEBUG:root:Unhandled type_constraint: Union[]
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.6_sdk:2.26.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.6_sdk:2.26.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f95dbda8bf8> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:16 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['external_9SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/Impulse\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper)\n  SqlTransform(beam:external:java:sql:v1)/BeamValuesRel_13/Create.Values/Read(CreateSource)/ParDo(BoundedSourceAsSDFWrapper)/ParMultiDo(BoundedSourceAsSDFWrapper):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'external_9SqlTransform(beam:external:java:sql:v1)/BeamZetaSqlCalcRel_17/ParDo(Calc)/ParMultiDo(Calc)\n  SqlTransform(beam:external:java:sql:v1)/BeamZetaSqlCalcRel_17/ParDo(Calc)/ParMultiDo(Calc):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_5\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2956>)_6\n  assert_that/Create/FlatMap(<lambda at core.py:2956>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_8\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_9\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_10\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_12\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_13\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_14\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_15\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_16\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_17\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_18\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'flink_master' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'pubsub_root_url' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok

----------------------------------------------------------------------
XML: nosetests-xlangSqlValidateRunner.xml
----------------------------------------------------------------------
XML: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 9 tests in 232.998s

OK

> Task :runners:flink:1.10:job-server:flinkJobServerCleanup

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':runners:flink:1.10:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 43m 10s
159 actionable tasks: 134 executed, 21 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/7dtc2asydon6u

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
No JDK named ‘JDK 1.8 (latest)’ found


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org