You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2022/01/31 00:48:32 UTC

Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1240

See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1240/display/redirect>

Changes:


------------------------------------------
[...truncated 619.57 KB...]
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001)\314\003\360R\254\355\000\005sr\000-org.apache.beam.sdk.io.kafka.KafkaRecordCoderI\320\251Y\032\341{#\002\000\001L\000\007kvCodert\000$Lorg/a\005F\000/\001F\024/sdk/c\0012\010s/K\t&\020;xr\000*Nm\000\t'$.Structure\tj8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\000C\001g\034C\335\325\211\256\274~\370\001/\020psr\000\"j1\000\r\221@j\000\275\271\035\243\3126\002\000\002L\000\010key\005I\010t\000\"n\315\000\005$ ;L\000\nvalue\005\016\024q\000~\000\006x\001\006\020\002sr\000)jy\000\034ByteArra\tj\034\351\277\005|\263\3746\333\001\263\010r\000&j8\000\024Atomic\005mT\307\354\265\314\205tPF\002\000\000xq\000~\000\002q\000~\000\n"
      >
    >
  >
  coders: <
    key: "gonWMPiLWTKvCoder"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "gonWMPiLWTByteArrayCoder"
      component_coder_ids: "gonWMPiLWTByteArrayCoder"
    >
  >
  coders: <
    key: "gonWMPiLWTKvCoder1"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "gonWMPiLWTSerializableCoder"
      component_coder_ids: "gonWMPiLWTVoidCoder"
    >
  >
  coders: <
    key: "gonWMPiLWTKvCoder2"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "gonWMPiLWTVarIntCoder"
      component_coder_ids: "gonWMPiLWTSerializableCoder1"
    >
  >
  coders: <
    key: "gonWMPiLWTKvCoder3"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "gonWMPiLWTVarIntCoder"
      component_coder_ids: "gonWMPiLWTTimestampedValueCoder"
    >
  >
  coders: <
    key: "gonWMPiLWTKvCoder4"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "gonWMPiLWTVarIntCoder"
      component_coder_ids: "gonWMPiLWTIterableCoder"
    >
  >
  coders: <
    key: "gonWMPiLWTSerializableCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\241\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000$N1\000\\io.BoundedSource!`v\023\177\370QA\0015\010r\000\035Z3\000\t,0\020\213\372\020Y\214\347\323\002\000\000xp"
      >
    >
  >
  coders: <
    key: "gonWMPiLWTSerializableCoder1"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\214\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000;N1\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "gonWMPiLWTTimestampedValueCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001?\315\003\270\254\355\000\005sr\000Aorg.apache.beam.sdk.values.TimestampedV\001\022\000$>\021\000LCoder\364\376\202p\274iX-\002\000\001L\000\nv\001)\005\030\034t\000\"Lorg/\t]\000/\001],/sdk/coders/\005$\020;xr\000*N\202\000\t%(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000,j1\000(Serializabl\t\277\034\010\364 \216\305ynl\005\327\\\004typet\000\021Ljava/lang/Class\001\300\000&jV\000\024Custom\005\215Pj\260\010\235\013;\035\013\002\000\000xq\000~\000\003vr\000;N;\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "gonWMPiLWTValueWithRecordIdCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001h\364\004\320\254\355\000\005sr\000Corg.apache.beam.sdk.values.ValueWithRecordId$B\022\000LCoder{\343\376bQ\014\204>\002\000\002L\000\007i\t\025\034t\000+Lorg/\t\\\000/\001\\\024/sdk/c\0012(s/ByteArray\005B\014;L\000\n\005u\005\016\010t\000\"n;\000\005$\020;xr\000*N\274\000\t`(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000)j1\0006\312\000\034\351\277\005|\263\3746\333\001:\010r\000&j8\000\024Atomic\005o\034\307\354\265\314\205tPF\0015 q\000~\000\003sr\000-N;\000$io.kafka.K\001\006\000R%\227\005B<I\320\251Y\032\341{#\002\000\001L\000\007kv\005\025\010t\000$nH\001\000K\t&\000;\025s\000\"Ns\000-P\r- j\000\275\271\035\243\3126\002!\353\010\010ke)\277\001\264\000\0022\303\001\005\022\rY$q\000~\000\006q\000~\000\016"
      >
    >
  >
  coders: <
    key: "gonWMPiLWTVarIntCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\330\001\350\254\355\000\005sr\000&org.apache.beam.sdk.coders.VarIntCoder\300K2\332\251KVh\002\000\000xrr5\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*jj\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  coders: <
    key: "gonWMPiLWTVoidCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\326\001\350\254\355\000\005sr\000$org.apache.beam.sdk.coders.VoidCoder\271\277U\233\350\r\257U\002\000\000xr\000&j3\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*j5\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
  environments: <
    key: "gonWMPiLWTbeam:env:docker:v1"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:timer:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:custom_window:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:param_windowed_value:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:sharded_key:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:harness_monitoring_infos:v1"
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:protocol:state_caching:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:deferred:v1"
        type_payload: "\n$b5649eb1-1080-4072-a85b-181b32417fef"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2022/01/31 00:27:54 Prepared job with id: go-job-2-1643588874540856204_4b88fd7a-afac-44ef-b9a9-be2fed1a6bfc and staging token: go-job-2-1643588874540856204_4b88fd7a-afac-44ef-b9a9-be2fed1a6bfc
2022/01/31 00:27:54 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1643588874550085463
    ptest.go:101: Failed to execute job: 	staging artifacts
        failed to stage /tmp/worker-2-1643588874550085463 in 3 attempts: failed to send chunks for /tmp/worker-2-1643588874550085463
        	caused by:
        chunk send failed
        	caused by:
        EOF; failed to send chunks for /tmp/worker-2-1643588874550085463
        	caused by:
        chunk send failed
        	caused by:
        EOF; failed to send chunks for /tmp/worker-2-1643588874550085463
        	caused by:
        chunk send failed
        	caused by:
        EOF; failed to send chunks for /tmp/worker-2-1643588874550085463
        	caused by:
        chunk send failed
        	caused by:
        EOF
--- FAIL: TestKafkaIO_BasicReadWrite (27.34s)
FAIL
FAIL	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	30.384s
FAIL

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 31m 40s
217 actionable tasks: 150 executed, 58 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/ohqvh7nzoim3u

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_XVR_Spark3 #1250

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1250/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1249

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1249/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13574] Large Wordcount (#16455)

[noreply] [BEAM-13293] Refactor JDBC IO Go Wrapper (#16686)


------------------------------------------
[...truncated 490.43 KB...]
You should consider upgrading via the '/usr/local/bin/python -m pip install --upgrade pip' command.
Removing intermediate container b42ee13d6754
 ---> a2a54303bd48
Step 14/31 : RUN pip check || (echo "Container does not include required Beam dependencies or has conflicting dependencies. If Beam dependencies have changed, you need to regenerate base_image_requirements.txt files. See: https://s.apache.org/beam-python-requirements-generate" && exit 1)
 ---> Running in 07e9286572f6
No broken requirements found.
Removing intermediate container 07e9286572f6
 ---> 592f109bb5ea
Step 15/31 : RUN pip freeze --all
 ---> Running in 807846b13f33
absl-py==0.15.0
apache-beam @ file:///opt/apache/beam/tars/apache-beam.tar.gz
astunparse==1.6.3
atomicwrites==1.4.0
attrs==21.4.0
beautifulsoup4==4.10.0
bs4==0.0.1
cached-property==1.5.2
cachetools==4.2.4
certifi==2021.10.8
charset-normalizer==2.0.10
clang==5.0
click==8.0.3
cloudpickle==2.0.0
crcmod==1.7
Cython==0.29.26
dataclasses==0.8
deprecation==2.1.0
dill==0.3.1.1
docker==5.0.3
docopt==0.6.2
execnet==1.9.0
fastavro==1.4.7
fasteners==0.16.3
flatbuffers==1.12
freezegun==1.1.0
future==0.18.2
gast==0.4.0
google-api-core==1.31.5
google-api-python-client==2.33.0
google-apitools==0.5.31
google-auth==1.35.0
google-auth-httplib2==0.1.0
google-auth-oauthlib==0.4.6
google-cloud-bigquery==2.31.0
google-cloud-bigquery-storage==2.10.1
google-cloud-bigtable==1.7.0
google-cloud-core==1.7.2
google-cloud-datastore==1.15.3
google-cloud-dlp==3.4.0
google-cloud-language==1.3.0
google-cloud-profiler==3.0.6
google-cloud-pubsub==2.9.0
google-cloud-pubsublite==1.2.0
google-cloud-recommendations-ai==0.2.0
google-cloud-spanner==1.19.1
google-cloud-videointelligence==1.16.1
google-cloud-vision==1.0.0
google-crc32c==1.3.0
google-pasta==0.2.0
google-python-cloud-debugger==2.18
google-resumable-media==2.1.0
googleapis-common-protos==1.54.0
greenlet==1.1.2
grpc-google-iam-v1==0.12.3
grpcio==1.43.0
grpcio-gcp==0.2.2
grpcio-status==1.43.0
guppy3==3.1.2
h5py==3.1.0
hdfs==2.6.0
httplib2==0.19.1
idna==3.3
importlib-metadata==4.8.3
joblib==1.1.0
keras==2.6.0
Keras-Preprocessing==1.1.2
libcst==0.3.23
Markdown==3.3.6
mmh3==3.0.0
mock==2.0.0
more-itertools==8.12.0
mypy-extensions==0.4.3
nltk==3.6.7
nose==1.3.7
numpy==1.19.5
oauth2client==4.1.3
oauthlib==3.1.1
opt-einsum==3.3.0
orjson==3.6.1
overrides==6.1.0
packaging==21.3
pandas==1.1.5
parameterized==0.7.5
pbr==5.8.0
pip==21.2.4
pluggy==0.13.1
proto-plus==1.19.8
protobuf==3.19.1
psycopg2-binary==2.9.3
py==1.11.0
pyarrow==6.0.1
pyasn1==0.4.8
pyasn1-modules==0.2.8
pydot==1.4.2
PyHamcrest==1.10.1
pymongo==3.12.3
pyparsing==2.4.7
pytest==4.6.11
pytest-forked==1.4.0
pytest-timeout==1.4.2
pytest-xdist==1.34.0
python-dateutil==2.8.2
python-snappy==0.6.0
pytz==2021.3
PyYAML==6.0
regex==2021.11.10
requests==2.27.0
requests-mock==1.9.3
requests-oauthlib==1.3.0
rsa==4.8
scikit-learn==0.24.2
scipy==1.5.4
setuptools==57.5.0
six==1.15.0
soupsieve==2.3.1
SQLAlchemy==1.4.29
tenacity==5.1.5
tensorboard==2.6.0
tensorboard-data-server==0.6.1
tensorboard-plugin-wit==1.8.0
tensorflow==2.6.2
tensorflow-estimator==2.6.0
termcolor==1.1.0
testcontainers==3.4.2
threadpoolctl==3.0.0
tqdm==4.62.3
typing-extensions==3.7.4.3
typing-inspect==0.7.1
typing-utils==0.1.0
uritemplate==4.1.1
urllib3==1.26.7
wcwidth==0.2.5
websocket-client==1.2.3
Werkzeug==2.0.2
wheel==0.37.0
wrapt==1.12.1
zipp==3.6.0
Removing intermediate container 807846b13f33
 ---> 1d5f93151c49
Step 16/31 : COPY target/LICENSE /opt/apache/beam/
 ---> ab75ba7b4fde
Step 17/31 : COPY target/LICENSE.python /opt/apache/beam/
 ---> b2bffd40c3c9
Step 18/31 : COPY target/NOTICE /opt/apache/beam/
 ---> eeb9502bcc2f
Step 19/31 : COPY target/launcher/linux_amd64/boot /opt/apache/beam/
 ---> cce566803da9
Step 20/31 : ENTRYPOINT ["/opt/apache/beam/boot"]
 ---> Running in 47cfadfa08de
Removing intermediate container 47cfadfa08de
 ---> 15ff2c6d85e5
Step 21/31 : FROM beam as third_party_licenses
 ---> 15ff2c6d85e5
Step 22/31 : ARG pull_licenses
 ---> Running in 8bfc6bd27c2e
Removing intermediate container 8bfc6bd27c2e
 ---> be725a4b684f
Step 23/31 : COPY target/license_scripts /tmp/license_scripts/
 ---> 31341b681e8d
Step 24/31 : COPY target/LICENSE target/go-licenses/* /opt/apache/beam/third_party_licenses/golang/
 ---> 032d6f175128
Step 25/31 : RUN rm /opt/apache/beam/third_party_licenses/golang/LICENSE
 ---> Running in ae4dc369a009
Removing intermediate container ae4dc369a009
 ---> 24d834d6965f
Step 26/31 : COPY target/license_scripts /tmp/license_scripts/
 ---> 91660cab5151
Step 27/31 : RUN if [ "$pull_licenses" = "true" ] ; then       pip install 'pip-licenses<3.0.0' pyyaml tenacity &&       python /tmp/license_scripts/pull_licenses_py.py ;     fi
 ---> Running in eadf22d1ea5f
Collecting pip-licenses<3.0.0
  Downloading pip_licenses-2.3.0-py3-none-any.whl (14 kB)
Requirement already satisfied: pyyaml in /usr/local/lib/python3.6/site-packages (6.0)
Requirement already satisfied: tenacity in /usr/local/lib/python3.6/site-packages (5.1.5)
Collecting PTable
  Downloading PTable-0.9.2.tar.gz (31 kB)
Requirement already satisfied: six>=1.9.0 in /usr/local/lib/python3.6/site-packages (from tenacity) (1.15.0)
Building wheels for collected packages: PTable
  Building wheel for PTable (setup.py): started
  Building wheel for PTable (setup.py): finished with status 'done'
  Created wheel for PTable: filename=PTable-0.9.2-py3-none-any.whl size=22924 sha256=e9df81cff94e4fc423e72fa594cfdfa7784e3561c492db8a3e86c44286d0772e
  Stored in directory: /root/.cache/pip/wheels/f3/65/67/71f473ec87ea4554d1bdfb1b5128cfe1414c2b113b72a1ee8e
Successfully built PTable
Installing collected packages: PTable, pip-licenses
Successfully installed PTable-0.9.2 pip-licenses-2.3.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
WARNING: You are using pip version 21.2.4; however, version 21.3.1 is available.
You should consider upgrading via the '/usr/local/bin/python -m pip install --upgrade pip' command.
INFO:root:Successfully pulled licenses for 134 dependencies
Skip pulling license for  bs4
Removing intermediate container eadf22d1ea5f
 ---> 72d46c9647ac
Step 28/31 : FROM beam
 ---> 15ff2c6d85e5
Step 29/31 : ARG pull_licenses
 ---> Running in 96163afafb59
Removing intermediate container 96163afafb59
 ---> a0f954f91eeb
Step 30/31 : COPY --from=third_party_licenses /opt/apache/beam/third_party_licenses /opt/apache/beam/third_party_licenses
 ---> 4d02e27e69b7
Step 31/31 : RUN if [ "$pull_licenses" != "true" ] ; then       rm -rf /opt/apache/beam/third_party_licenses ;     fi
 ---> Running in 9c4619d17719
Removing intermediate container 9c4619d17719
 ---> 589a6e169e2a
Successfully built 589a6e169e2a
Successfully tagged apache/beam_python3.6_sdk:2.37.0.dev

> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 15m 12s
205 actionable tasks: 139 executed, 57 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/pqwcm7ucelnji

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1248

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1248/display/redirect?page=changes>

Changes:

[career] [BEAM-13734] Support cache directories that use GCS buckets

[noreply] Merge pull request #16655 from [BEAM-12164]: Add retry protection to

[noreply] Merge pull request #16586 from [BEAM-13731] FhirIO: Add support for

[noreply] [BEAM-13011] Adds a link to Multi-language Pipelines Tips wiki page

[noreply] [BEAM-12572] Run python examples on multiple runners (#16154)


------------------------------------------
[...truncated 483.88 KB...]
Removing intermediate container a2bb51c3c428
 ---> 9a77ceeec764
Step 14/31 : RUN pip check || (echo "Container does not include required Beam dependencies or has conflicting dependencies. If Beam dependencies have changed, you need to regenerate base_image_requirements.txt files. See: https://s.apache.org/beam-python-requirements-generate" && exit 1)
 ---> Running in 4a0b73b90111
No broken requirements found.
Removing intermediate container 4a0b73b90111
 ---> 7aacfc828cca
Step 15/31 : RUN pip freeze --all
 ---> Running in e75a0cfa724d
absl-py==0.15.0
apache-beam @ file:///opt/apache/beam/tars/apache-beam.tar.gz
astunparse==1.6.3
atomicwrites==1.4.0
attrs==21.4.0
beautifulsoup4==4.10.0
bs4==0.0.1
cached-property==1.5.2
cachetools==4.2.4
certifi==2021.10.8
charset-normalizer==2.0.10
clang==5.0
click==8.0.3
cloudpickle==2.0.0
crcmod==1.7
Cython==0.29.26
dataclasses==0.8
deprecation==2.1.0
dill==0.3.1.1
docker==5.0.3
docopt==0.6.2
execnet==1.9.0
fastavro==1.4.7
fasteners==0.16.3
flatbuffers==1.12
freezegun==1.1.0
future==0.18.2
gast==0.4.0
google-api-core==1.31.5
google-api-python-client==2.33.0
google-apitools==0.5.31
google-auth==1.35.0
google-auth-httplib2==0.1.0
google-auth-oauthlib==0.4.6
google-cloud-bigquery==2.31.0
google-cloud-bigquery-storage==2.10.1
google-cloud-bigtable==1.7.0
google-cloud-core==1.7.2
google-cloud-datastore==1.15.3
google-cloud-dlp==3.4.0
google-cloud-language==1.3.0
google-cloud-profiler==3.0.6
google-cloud-pubsub==2.9.0
google-cloud-pubsublite==1.2.0
google-cloud-recommendations-ai==0.2.0
google-cloud-spanner==1.19.1
google-cloud-videointelligence==1.16.1
google-cloud-vision==1.0.0
google-crc32c==1.3.0
google-pasta==0.2.0
google-python-cloud-debugger==2.18
google-resumable-media==2.1.0
googleapis-common-protos==1.54.0
greenlet==1.1.2
grpc-google-iam-v1==0.12.3
grpcio==1.43.0
grpcio-gcp==0.2.2
grpcio-status==1.43.0
guppy3==3.1.2
h5py==3.1.0
hdfs==2.6.0
httplib2==0.19.1
idna==3.3
importlib-metadata==4.8.3
joblib==1.1.0
keras==2.6.0
Keras-Preprocessing==1.1.2
libcst==0.3.23
Markdown==3.3.6
mmh3==3.0.0
mock==2.0.0
more-itertools==8.12.0
mypy-extensions==0.4.3
nltk==3.6.7
nose==1.3.7
numpy==1.19.5
oauth2client==4.1.3
oauthlib==3.1.1
opt-einsum==3.3.0
orjson==3.6.1
overrides==6.1.0
packaging==21.3
pandas==1.1.5
parameterized==0.7.5
pbr==5.8.0
pip==21.2.4
pluggy==0.13.1
proto-plus==1.19.8
protobuf==3.19.1
psycopg2-binary==2.9.3
py==1.11.0
pyarrow==6.0.1
pyasn1==0.4.8
pyasn1-modules==0.2.8
pydot==1.4.2
PyHamcrest==1.10.1
pymongo==3.12.3
pyparsing==2.4.7
pytest==4.6.11
pytest-forked==1.4.0
pytest-timeout==1.4.2
pytest-xdist==1.34.0
python-dateutil==2.8.2
python-snappy==0.6.0
pytz==2021.3
PyYAML==6.0
regex==2021.11.10
requests==2.27.0
requests-mock==1.9.3
requests-oauthlib==1.3.0
rsa==4.8
scikit-learn==0.24.2
scipy==1.5.4
setuptools==57.5.0
six==1.15.0
soupsieve==2.3.1
SQLAlchemy==1.4.29
tenacity==5.1.5
tensorboard==2.6.0
tensorboard-data-server==0.6.1
tensorboard-plugin-wit==1.8.0
tensorflow==2.6.2
tensorflow-estimator==2.6.0
termcolor==1.1.0
testcontainers==3.4.2
threadpoolctl==3.0.0
tqdm==4.62.3
typing-extensions==3.7.4.3
typing-inspect==0.7.1
typing-utils==0.1.0
uritemplate==4.1.1
urllib3==1.26.7
wcwidth==0.2.5
websocket-client==1.2.3
Werkzeug==2.0.2
wheel==0.37.0
wrapt==1.12.1
zipp==3.6.0
Removing intermediate container e75a0cfa724d
 ---> 86dedca8101d
Step 16/31 : COPY target/LICENSE /opt/apache/beam/
 ---> 4db5af48d17b
Step 17/31 : COPY target/LICENSE.python /opt/apache/beam/
 ---> cbc5e70c2a84
Step 18/31 : COPY target/NOTICE /opt/apache/beam/
 ---> 40bead9c8272
Step 19/31 : COPY target/launcher/linux_amd64/boot /opt/apache/beam/
 ---> 966d461459be
Step 20/31 : ENTRYPOINT ["/opt/apache/beam/boot"]
 ---> Running in 0d26b9d9bde9
Removing intermediate container 0d26b9d9bde9
 ---> 87012fb636f0
Step 21/31 : FROM beam as third_party_licenses
 ---> 87012fb636f0
Step 22/31 : ARG pull_licenses
 ---> Running in 782f165cc1a7
Removing intermediate container 782f165cc1a7
 ---> 5c1c243da0e5
Step 23/31 : COPY target/license_scripts /tmp/license_scripts/
 ---> 57979905c50c
Step 24/31 : COPY target/LICENSE target/go-licenses/* /opt/apache/beam/third_party_licenses/golang/
 ---> eb69ced9f6f6
Step 25/31 : RUN rm /opt/apache/beam/third_party_licenses/golang/LICENSE
 ---> Running in bf95121cf711
Removing intermediate container bf95121cf711
 ---> 633d8fa8422e
Step 26/31 : COPY target/license_scripts /tmp/license_scripts/
 ---> b1365d461fd5
Step 27/31 : RUN if [ "$pull_licenses" = "true" ] ; then       pip install 'pip-licenses<3.0.0' pyyaml tenacity &&       python /tmp/license_scripts/pull_licenses_py.py ;     fi
 ---> Running in 0841af468c62
Collecting pip-licenses<3.0.0
  Downloading pip_licenses-2.3.0-py3-none-any.whl (14 kB)
Requirement already satisfied: pyyaml in /usr/local/lib/python3.6/site-packages (6.0)
Requirement already satisfied: tenacity in /usr/local/lib/python3.6/site-packages (5.1.5)
Collecting PTable
  Downloading PTable-0.9.2.tar.gz (31 kB)
Requirement already satisfied: six>=1.9.0 in /usr/local/lib/python3.6/site-packages (from tenacity) (1.15.0)
Building wheels for collected packages: PTable
  Building wheel for PTable (setup.py): started
  Building wheel for PTable (setup.py): finished with status 'done'
  Created wheel for PTable: filename=PTable-0.9.2-py3-none-any.whl size=22924 sha256=a06506abc7f3ecbb80525d60bcb634a00eaaf74df6146f23804a6f7352252aa6
  Stored in directory: /root/.cache/pip/wheels/f3/65/67/71f473ec87ea4554d1bdfb1b5128cfe1414c2b113b72a1ee8e
Successfully built PTable
Installing collected packages: PTable, pip-licenses
Successfully installed PTable-0.9.2 pip-licenses-2.3.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
WARNING: You are using pip version 21.2.4; however, version 21.3.1 is available.
You should consider upgrading via the '/usr/local/bin/python -m pip install --upgrade pip' command.
INFO:root:Successfully pulled licenses for 134 dependencies
Skip pulling license for  bs4
Removing intermediate container 0841af468c62
 ---> 096331eb6aef
Step 28/31 : FROM beam
 ---> 87012fb636f0
Step 29/31 : ARG pull_licenses
 ---> Running in bb13ca51916f
Removing intermediate container bb13ca51916f
 ---> ce286a8dabd2
Step 30/31 : COPY --from=third_party_licenses /opt/apache/beam/third_party_licenses /opt/apache/beam/third_party_licenses
 ---> ec19c1a554e5
Step 31/31 : RUN if [ "$pull_licenses" != "true" ] ; then       rm -rf /opt/apache/beam/third_party_licenses ;     fi
 ---> Running in 6bf9ac716186
Removing intermediate container 6bf9ac716186
 ---> bf72763aa7ec
Successfully built bf72763aa7ec
Successfully tagged apache/beam_python3.6_sdk:2.37.0.dev

> Task :runners:spark:3:job-server:sparkJobServerCleanup
Stopping job server pid: 730153.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 17m 29s
205 actionable tasks: 140 executed, 56 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/pxpfu5nptve6o

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1247

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1247/display/redirect>

Changes:


------------------------------------------
[...truncated 726.38 KB...]
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7f3af45364e0>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Waiting for grpc channel to be ready at localhost:41441.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:subprocess_server.py:98 Waiting for grpc channel to be ready at localhost:41441.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
__________________ SqlTransformTest.test_windowing_before_sql __________________

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_windowing_before_sql>

    def test_windowing_before_sql(self):
      with TestPipeline() as p:
        out = (
            p | beam.Create([
                SimpleRow(5, "foo", 1.),
                SimpleRow(15, "bar", 2.),
                SimpleRow(25, "baz", 3.)
            ])
            | beam.Map(lambda v: beam.window.TimestampedValue(v, v.id)).
            with_output_types(SimpleRow)
            | beam.WindowInto(
                beam.window.FixedWindows(10)).with_output_types(SimpleRow)
            | SqlTransform("SELECT COUNT(*) as `count` FROM PCOLLECTION"))
>       assert_that(out, equal_to([(1, ), (1, ), (1, )]))

apache_beam/transforms/sql_test.py:183: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7f3afbccd630>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Waiting for grpc channel to be ready at localhost:46071.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:subprocess_server.py:98 Waiting for grpc channel to be ready at localhost:46071.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
_________________ SqlTransformTest.test_zetasql_generate_data __________________

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_zetasql_generate_data>

    def test_zetasql_generate_data(self):
      with TestPipeline() as p:
        out = p | SqlTransform(
            """SELECT
              CAST(1 AS INT64) AS `int`,
              CAST('foo' AS STRING) AS `str`,
              CAST(3.14  AS FLOAT64) AS `flt`""",
            dialect="zetasql")
>       assert_that(out, equal_to([(1, "foo", 3.14)]))

apache_beam/transforms/sql_test.py:168: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7f3af45c01d0>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Waiting for grpc channel to be ready at localhost:48589.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:subprocess_server.py:98 Waiting for grpc channel to be ready at localhost:48589.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/python/pytest_xlangSqlValidateRunner.xml> -
================= 9 failed, 5170 deselected in 712.11 seconds ==================

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build completed with 4 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 44m 49s
217 actionable tasks: 151 executed, 57 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/chwbn2srftqos

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1246

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1246/display/redirect>

Changes:


------------------------------------------
[...truncated 263.64 KB...]
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\214\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000;N1\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "oItIhrCYmaTimestampedValueCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001?\315\003\270\254\355\000\005sr\000Aorg.apache.beam.sdk.values.TimestampedV\001\022\000$>\021\000LCoder\364\376\202p\274iX-\002\000\001L\000\nv\001)\005\030\034t\000\"Lorg/\t]\000/\001],/sdk/coders/\005$\020;xr\000*N\202\000\t%(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000,j1\000(Serializabl\t\277\034\010\364 \216\305ynl\005\327\\\004typet\000\021Ljava/lang/Class\001\300\000&jV\000\024Custom\005\215Pj\260\010\235\013;\035\013\002\000\000xq\000~\000\003vr\000;N;\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "oItIhrCYmaValueWithRecordIdCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001h\364\004\320\254\355\000\005sr\000Corg.apache.beam.sdk.values.ValueWithRecordId$B\022\000LCoder{\343\376bQ\014\204>\002\000\002L\000\007i\t\025\034t\000+Lorg/\t\\\000/\001\\\024/sdk/c\0012(s/ByteArray\005B\014;L\000\n\005u\005\016\010t\000\"n;\000\005$\020;xr\000*N\274\000\t`(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000)j1\0006\312\000\034\351\277\005|\263\3746\333\001:\010r\000&j8\000\024Atomic\005o\034\307\354\265\314\205tPF\0015 q\000~\000\003sr\000-N;\000$io.kafka.K\001\006\000R%\227\005B<I\320\251Y\032\341{#\002\000\001L\000\007kv\005\025\010t\000$nH\001\000K\t&\000;\025s\000\"Ns\000-P\r- j\000\275\271\035\243\3126\002!\353\010\010ke)\277\001\264\000\0022\303\001\005\022\rY$q\000~\000\006q\000~\000\016"
      >
    >
  >
  coders: <
    key: "oItIhrCYmaVarIntCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\330\001\350\254\355\000\005sr\000&org.apache.beam.sdk.coders.VarIntCoder\300K2\332\251KVh\002\000\000xrr5\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*jj\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  coders: <
    key: "oItIhrCYmaVoidCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\326\001\350\254\355\000\005sr\000$org.apache.beam.sdk.coders.VoidCoder\271\277U\233\350\r\257U\002\000\000xr\000&j3\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*j5\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
  environments: <
    key: "oItIhrCYmabeam:env:docker:v1"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:timer:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:custom_window:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:param_windowed_value:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:sharded_key:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:harness_monitoring_infos:v1"
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:protocol:state_caching:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nL/tmp/artifacts/icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar\022@79bbed34591f15783869a605b839f02a9c034a3cec699aa5aafe622b13d3afe5"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n=icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar\022@50b15308fb1be9c2d86747f5046d4541f73398735a642c7ca655ee4432a906a1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nI/tmp/artifacts/localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar\022@7541eac86c5a4d50a37d323c31c90f61aad9dff9b07fadae764c5a1e2d5a2a7b"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n:localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar\022@5dd5277501aba1738ff4d0ac7c84e904445871b6c65d51cb8dbbcd5970a1fb70"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nG/tmp/artifacts/cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar\022@62acee297d509cb08ea347308ca45d0611ab8a9fe5b4825983e00f4fad2d50f0"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n8cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nD/tmp/artifacts/dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar\022@761129d7ceae74417a5fa72166eb3e449cd6466ce571cc71d7f6ed95759ac952"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-fxxlqTpKpNpL9ctKaFbujqr7izb4QZD1NumRcxv0H8Q.jar\022@7f1c65a93a4aa4da4bf5cb4a6856ee8eaafb8b36f84190f536e991731bf41fc4"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\ncbeam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-fxxlqTpKpNpL9ctKaFbujqr7izb4QZD1NumRcxv0H8Q.jar"
      >
    >
  >
>
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s1"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2022/02/01 12:49:43 Prepared job with id: go-job-2-1643719783626165959_847ed244-6317-4446-b8ac-0d333c764e07 and staging token: go-job-2-1643719783626165959_847ed244-6317-4446-b8ac-0d333c764e07
2022/02/01 12:49:43 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1643719783637503019
2022/02/01 12:49:45 Staged binary artifact with token: 
2022/02/01 12:49:45 Submitted job: go0job0201643719783626165959-jenkins-0201124945-e5378b61_7837da38-092d-4cdd-9fa0-8b5991881e2c
2022/02/01 12:49:45 Job state: STOPPED
2022/02/01 12:49:45 Job state: STARTING
2022/02/01 12:49:45 Job state: RUNNING
2022/02/01 12:50:14 Job state: DONE
2022/02/01 12:50:14 Warning: 6 errors during metrics processing: [failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"oItIhrCYmaExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"oItIhrCYmaExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xbcA\xbcA\xbcA"  labels:{key:"PCOLLECTION"  value:"oItIhrCYmaExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"oItIhrCYmaExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"oItIhrCYmaExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"oItIhrCYmaExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}]
--- PASS: TestKafkaIO_BasicReadWrite (51.60s)
PASS
ok  	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	54.655s
FAIL

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build completed with 4 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 46m 8s
217 actionable tasks: 152 executed, 56 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/zlw5efcwvno3i

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1245

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1245/display/redirect?page=changes>

Changes:

[noreply] [BEAM-12240] Add Java 17 support (#16568)

[noreply] [BEAM-13760] Add random component to default python dataflow job name


------------------------------------------
[...truncated 686.49 KB...]
self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_tagged_join>

    def test_tagged_join(self):
      with TestPipeline() as p:
        enrich = (
            p | "Create enrich" >> beam.Create(
                [Enrich(1, "a"), Enrich(2, "b"), Enrich(26, "z")]))
        simple = (
            p | "Create simple" >> beam.Create([
                SimpleRow(1, "foo", 3.14),
                SimpleRow(26, "bar", 1.11),
                SimpleRow(1, "baz", 2.34)
            ]))
        out = ({
            'simple': simple, 'enrich': enrich
        }
               | SqlTransform(
                   """
                SELECT
                  simple.`id` AS `id`,
                  enrich.metadata AS metadata
                FROM simple
                JOIN enrich
                ON simple.`id` = enrich.`id`"""))
>       assert_that(out, equal_to([(1, "a"), (26, "z"), (1, "a")]))

apache_beam/transforms/sql_test.py:149: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7f9c104caa20>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
__________________ SqlTransformTest.test_windowing_before_sql __________________

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_windowing_before_sql>

    def test_windowing_before_sql(self):
      with TestPipeline() as p:
        out = (
            p | beam.Create([
                SimpleRow(5, "foo", 1.),
                SimpleRow(15, "bar", 2.),
                SimpleRow(25, "baz", 3.)
            ])
            | beam.Map(lambda v: beam.window.TimestampedValue(v, v.id)).
            with_output_types(SimpleRow)
            | beam.WindowInto(
                beam.window.FixedWindows(10)).with_output_types(SimpleRow)
            | SqlTransform("SELECT COUNT(*) as `count` FROM PCOLLECTION"))
>       assert_that(out, equal_to([(1, ), (1, ), (1, )]))

apache_beam/transforms/sql_test.py:183: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7f9c1032fc18>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
_________________ SqlTransformTest.test_zetasql_generate_data __________________

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_zetasql_generate_data>

    def test_zetasql_generate_data(self):
      with TestPipeline() as p:
        out = p | SqlTransform(
            """SELECT
              CAST(1 AS INT64) AS `int`,
              CAST('foo' AS STRING) AS `str`,
              CAST(3.14  AS FLOAT64) AS `flt`""",
            dialect="zetasql")
>       assert_that(out, equal_to([(1, "foo", 3.14)]))

apache_beam/transforms/sql_test.py:168: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7f9c102d38d0>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/python/pytest_xlangSqlValidateRunner.xml> -
============ 7 failed, 2 passed, 5170 deselected in 514.20 seconds =============

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 37m 15s
217 actionable tasks: 152 executed, 56 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/aj5ie55pqn4oo

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1244

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1244/display/redirect?page=changes>

Changes:

[daria.malkova] Support SCIO SDK via sbt projects

[samuelw] [BEAM-11648] Share thread pool across RetryManager instances.

[Pablo Estrada] Exclude per-key order tests on Twister2 runner

[Heejong Lee] Fix Java SDK container image name for load-tests and nexmark

[noreply] Merge pull request #16493 from [BEAM-13632][Playground] Save catalog

[noreply] Exclude jul-to-slf4j from Spark runner in quickstart POM templates

[noreply] [BEAM-11936] Enable a few errorprone checks that were broken by pinned

[noreply] [BEAM-13780] Add CONTRIBUTING.md pointing to main guide (#16666)

[noreply] [BEAM-13777] Accept cache capacity as input parameter instead of default

[noreply] [BEAM-13051][A] Enable pylint warnings

[noreply] [BEAM-13779] Fix pr labeling (#16665)

[noreply] Merge pull request #16581 from [BEAM-12164]: Add

[noreply] Fix labeler trigger (#16674)

[noreply] [BEAM-13781] Exclude grpc-netty-shaded from gax-grpc's dependency

[noreply] [BEAM-13051] Fixed pylint warnings : raising-non-exception (E0710),

[noreply] [BEAM-13740] Correctly install go before running tests (#16673)

[noreply] [BEAM-12830] Update local Docker env Go version. (#16670)

[noreply] [BEAM-13051][B] Enable pylint warnings

[noreply] [BEAM-13430] Revert Spark libraries in Spark runner to provided (#16675)


------------------------------------------
[...truncated 255.84 KB...]
  >
  coders: <
    key: "sMvLGixwTJSerializableCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\241\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000$N1\000\\io.BoundedSource!`v\023\177\370QA\0015\010r\000\035Z3\000\t,0\020\213\372\020Y\214\347\323\002\000\000xp"
      >
    >
  >
  coders: <
    key: "sMvLGixwTJSerializableCoder1"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\214\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000;N1\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "sMvLGixwTJTimestampedValueCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001?\315\003\270\254\355\000\005sr\000Aorg.apache.beam.sdk.values.TimestampedV\001\022\000$>\021\000LCoder\364\376\202p\274iX-\002\000\001L\000\nv\001)\005\030\034t\000\"Lorg/\t]\000/\001],/sdk/coders/\005$\020;xr\000*N\202\000\t%(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000,j1\000(Serializabl\t\277\034\010\364 \216\305ynl\005\327\\\004typet\000\021Ljava/lang/Class\001\300\000&jV\000\024Custom\005\215Pj\260\010\235\013;\035\013\002\000\000xq\000~\000\003vr\000;N;\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "sMvLGixwTJValueWithRecordIdCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001h\364\004\320\254\355\000\005sr\000Corg.apache.beam.sdk.values.ValueWithRecordId$B\022\000LCoder{\343\376bQ\014\204>\002\000\002L\000\007i\t\025\034t\000+Lorg/\t\\\000/\001\\\024/sdk/c\0012(s/ByteArray\005B\014;L\000\n\005u\005\016\010t\000\"n;\000\005$\020;xr\000*N\274\000\t`(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000)j1\0006\312\000\034\351\277\005|\263\3746\333\001:\010r\000&j8\000\024Atomic\005o\034\307\354\265\314\205tPF\0015 q\000~\000\003sr\000-N;\000$io.kafka.K\001\006\000R%\227\005B<I\320\251Y\032\341{#\002\000\001L\000\007kv\005\025\010t\000$nH\001\000K\t&\000;\025s\000\"Ns\000-P\r- j\000\275\271\035\243\3126\002!\353\010\010ke)\277\001\264\000\0022\303\001\005\022\rY$q\000~\000\006q\000~\000\016"
      >
    >
  >
  coders: <
    key: "sMvLGixwTJVarIntCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\330\001\350\254\355\000\005sr\000&org.apache.beam.sdk.coders.VarIntCoder\300K2\332\251KVh\002\000\000xrr5\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*jj\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  coders: <
    key: "sMvLGixwTJVoidCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\326\001\350\254\355\000\005sr\000$org.apache.beam.sdk.coders.VoidCoder\271\277U\233\350\r\257U\002\000\000xr\000&j3\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*j5\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
  environments: <
    key: "sMvLGixwTJbeam:env:docker:v1"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:timer:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:custom_window:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:param_windowed_value:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:sharded_key:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:harness_monitoring_infos:v1"
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:protocol:state_caching:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nL/tmp/artifacts/icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar\022@79bbed34591f15783869a605b839f02a9c034a3cec699aa5aafe622b13d3afe5"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n=icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar\022@50b15308fb1be9c2d86747f5046d4541f73398735a642c7ca655ee4432a906a1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nI/tmp/artifacts/localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar\022@7541eac86c5a4d50a37d323c31c90f61aad9dff9b07fadae764c5a1e2d5a2a7b"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n:localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar\022@5dd5277501aba1738ff4d0ac7c84e904445871b6c65d51cb8dbbcd5970a1fb70"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nG/tmp/artifacts/cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar\022@62acee297d509cb08ea347308ca45d0611ab8a9fe5b4825983e00f4fad2d50f0"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n8cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nD/tmp/artifacts/dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar\022@761129d7ceae74417a5fa72166eb3e449cd6466ce571cc71d7f6ed95759ac952"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-fxxlqTpKpNpL9ctKaFbujqr7izb4QZD1NumRcxv0H8Q.jar\022@7f1c65a93a4aa4da4bf5cb4a6856ee8eaafb8b36f84190f536e991731bf41fc4"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\ncbeam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-fxxlqTpKpNpL9ctKaFbujqr7izb4QZD1NumRcxv0H8Q.jar"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2022/02/01 01:04:50 Prepared job with id: go-job-2-1643677490432441732_d16c2175-85e2-40fb-9dfe-ed347f8af395 and staging token: go-job-2-1643677490432441732_d16c2175-85e2-40fb-9dfe-ed347f8af395
2022/02/01 01:04:50 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1643677490444280258
2022/02/01 01:04:52 Staged binary artifact with token: 
2022/02/01 01:04:52 Submitted job: go0job0201643677490432441732-jenkins-0201010452-3dc80ccd_c877be66-380e-412b-8655-f71353f39432
2022/02/01 01:04:52 Job state: STOPPED
2022/02/01 01:04:52 Job state: STARTING
2022/02/01 01:04:52 Job state: RUNNING
2022/02/01 01:05:23 Job state: DONE
2022/02/01 01:05:23 Warning: 6 errors during metrics processing: [failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"sMvLGixwTJExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"sMvLGixwTJExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"sMvLGixwTJExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"sMvLGixwTJExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xbcA\xbcA\xbcA"  labels:{key:"PCOLLECTION"  value:"sMvLGixwTJExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"sMvLGixwTJExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}]
--- PASS: TestKafkaIO_BasicReadWrite (51.78s)
PASS
ok  	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	54.800s
FAIL

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build completed with 3 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 42m 47s
217 actionable tasks: 152 executed, 56 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/ut3ja5qwwos4c

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1243

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1243/display/redirect?page=changes>

Changes:

[daria.malkova] Change executable name fo go tests

[avilovpavel6] Fix java test

[noreply] [BEAM-13769] Skip test_main_session_not_staged_when_using_cloudpickle

[noreply] [BEAM-6744] Support implicitly setting project id in Go Dataflow runner


------------------------------------------
[...truncated 620.03 KB...]
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\214\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000;N1\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "uhHkIdgugDTimestampedValueCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001?\315\003\270\254\355\000\005sr\000Aorg.apache.beam.sdk.values.TimestampedV\001\022\000$>\021\000LCoder\364\376\202p\274iX-\002\000\001L\000\nv\001)\005\030\034t\000\"Lorg/\t]\000/\001],/sdk/coders/\005$\020;xr\000*N\202\000\t%(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000,j1\000(Serializabl\t\277\034\010\364 \216\305ynl\005\327\\\004typet\000\021Ljava/lang/Class\001\300\000&jV\000\024Custom\005\215Pj\260\010\235\013;\035\013\002\000\000xq\000~\000\003vr\000;N;\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "uhHkIdgugDValueWithRecordIdCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001h\364\004\320\254\355\000\005sr\000Corg.apache.beam.sdk.values.ValueWithRecordId$B\022\000LCoder{\343\376bQ\014\204>\002\000\002L\000\007i\t\025\034t\000+Lorg/\t\\\000/\001\\\024/sdk/c\0012(s/ByteArray\005B\014;L\000\n\005u\005\016\010t\000\"n;\000\005$\020;xr\000*N\274\000\t`(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000)j1\0006\312\000\034\351\277\005|\263\3746\333\001:\010r\000&j8\000\024Atomic\005o\034\307\354\265\314\205tPF\0015 q\000~\000\003sr\000-N;\000$io.kafka.K\001\006\000R%\227\005B<I\320\251Y\032\341{#\002\000\001L\000\007kv\005\025\010t\000$nH\001\000K\t&\000;\025s\000\"Ns\000-P\r- j\000\275\271\035\243\3126\002!\353\010\010ke)\277\001\264\000\0022\303\001\005\022\rY$q\000~\000\006q\000~\000\016"
      >
    >
  >
  coders: <
    key: "uhHkIdgugDVarIntCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\330\001\350\254\355\000\005sr\000&org.apache.beam.sdk.coders.VarIntCoder\300K2\332\251KVh\002\000\000xrr5\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*jj\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  coders: <
    key: "uhHkIdgugDVoidCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\326\001\350\254\355\000\005sr\000$org.apache.beam.sdk.coders.VoidCoder\271\277U\233\350\r\257U\002\000\000xr\000&j3\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*j5\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
  environments: <
    key: "uhHkIdgugDbeam:env:docker:v1"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:timer:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:custom_window:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:param_windowed_value:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:sharded_key:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:harness_monitoring_infos:v1"
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:protocol:state_caching:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nL/tmp/artifacts/icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar\022@79bbed34591f15783869a605b839f02a9c034a3cec699aa5aafe622b13d3afe5"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n=icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar\022@50b15308fb1be9c2d86747f5046d4541f73398735a642c7ca655ee4432a906a1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nI/tmp/artifacts/localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar\022@7541eac86c5a4d50a37d323c31c90f61aad9dff9b07fadae764c5a1e2d5a2a7b"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n:localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar\022@5dd5277501aba1738ff4d0ac7c84e904445871b6c65d51cb8dbbcd5970a1fb70"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nG/tmp/artifacts/cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar\022@62acee297d509cb08ea347308ca45d0611ab8a9fe5b4825983e00f4fad2d50f0"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n8cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nD/tmp/artifacts/dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar\022@761129d7ceae74417a5fa72166eb3e449cd6466ce571cc71d7f6ed95759ac952"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-kTA9K1krhOJ9gdisYu_EQwLhuMbR6lpQm_JapiyjCrw.jar\022@91303d2b592b84e27d81d8ac62efc44302e1b8c6d1ea5a509bf25aa62ca30abc"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\ncbeam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-kTA9K1krhOJ9gdisYu_EQwLhuMbR6lpQm_JapiyjCrw.jar"
      >
    >
  >
>
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s1"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2022/01/31 18:35:33 Prepared job with id: go-job-2-1643654133864732416_7d2be888-df69-464f-9f2e-37b348937ad3 and staging token: go-job-2-1643654133864732416_7d2be888-df69-464f-9f2e-37b348937ad3
2022/01/31 18:35:33 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1643654133875444176
2022/01/31 18:35:36 Staged binary artifact with token: 
2022/01/31 18:35:36 Submitted job: go0job0201643654133864732416-jenkins-0131183536-e5bf7df3_98644303-b0dd-49a1-b8fc-2b6207577da5
2022/01/31 18:35:36 Job state: STOPPED
2022/01/31 18:35:36 Job state: STARTING
2022/01/31 18:35:36 Job state: RUNNING
2022/01/31 18:36:18 Job state: DONE
2022/01/31 18:36:18 Warning: 6 errors during metrics processing: [failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"uhHkIdgugDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"uhHkIdgugDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"uhHkIdgugDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"uhHkIdgugDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"uhHkIdgugDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xbcA\xbcA\xbcA"  labels:{key:"PCOLLECTION"  value:"uhHkIdgugDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"}]
--- PASS: TestKafkaIO_BasicReadWrite (64.18s)
PASS
ok  	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	67.205s
FAIL

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build completed with 4 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 46m 6s
217 actionable tasks: 151 executed, 57 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/4mdmxfeafzst2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1242

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1242/display/redirect?page=changes>

Changes:

[mrudary] Generalize S3FileSystem to support multiple URI schemes.


------------------------------------------
[...truncated 653.51 KB...]
  coders: <
    key: "zgRQRyxwxDKvCoder2"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "zgRQRyxwxDVarIntCoder"
      component_coder_ids: "zgRQRyxwxDSerializableCoder1"
    >
  >
  coders: <
    key: "zgRQRyxwxDKvCoder3"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "zgRQRyxwxDVarIntCoder"
      component_coder_ids: "zgRQRyxwxDTimestampedValueCoder"
    >
  >
  coders: <
    key: "zgRQRyxwxDKvCoder4"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "zgRQRyxwxDVarIntCoder"
      component_coder_ids: "zgRQRyxwxDIterableCoder"
    >
  >
  coders: <
    key: "zgRQRyxwxDSerializableCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\241\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000$N1\000\\io.BoundedSource!`v\023\177\370QA\0015\010r\000\035Z3\000\t,0\020\213\372\020Y\214\347\323\002\000\000xp"
      >
    >
  >
  coders: <
    key: "zgRQRyxwxDSerializableCoder1"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\214\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000;N1\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "zgRQRyxwxDTimestampedValueCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001?\315\003\270\254\355\000\005sr\000Aorg.apache.beam.sdk.values.TimestampedV\001\022\000$>\021\000LCoder\364\376\202p\274iX-\002\000\001L\000\nv\001)\005\030\034t\000\"Lorg/\t]\000/\001],/sdk/coders/\005$\020;xr\000*N\202\000\t%(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000,j1\000(Serializabl\t\277\034\010\364 \216\305ynl\005\327\\\004typet\000\021Ljava/lang/Class\001\300\000&jV\000\024Custom\005\215Pj\260\010\235\013;\035\013\002\000\000xq\000~\000\003vr\000;N;\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "zgRQRyxwxDValueWithRecordIdCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001h\364\004\320\254\355\000\005sr\000Corg.apache.beam.sdk.values.ValueWithRecordId$B\022\000LCoder{\343\376bQ\014\204>\002\000\002L\000\007i\t\025\034t\000+Lorg/\t\\\000/\001\\\024/sdk/c\0012(s/ByteArray\005B\014;L\000\n\005u\005\016\010t\000\"n;\000\005$\020;xr\000*N\274\000\t`(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000)j1\0006\312\000\034\351\277\005|\263\3746\333\001:\010r\000&j8\000\024Atomic\005o\034\307\354\265\314\205tPF\0015 q\000~\000\003sr\000-N;\000$io.kafka.K\001\006\000R%\227\005B<I\320\251Y\032\341{#\002\000\001L\000\007kv\005\025\010t\000$nH\001\000K\t&\000;\025s\000\"Ns\000-P\r- j\000\275\271\035\243\3126\002!\353\010\010ke)\277\001\264\000\0022\303\001\005\022\rY$q\000~\000\006q\000~\000\016"
      >
    >
  >
  coders: <
    key: "zgRQRyxwxDVarIntCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\330\001\350\254\355\000\005sr\000&org.apache.beam.sdk.coders.VarIntCoder\300K2\332\251KVh\002\000\000xrr5\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*jj\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  coders: <
    key: "zgRQRyxwxDVoidCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\326\001\350\254\355\000\005sr\000$org.apache.beam.sdk.coders.VoidCoder\271\277U\233\350\r\257U\002\000\000xr\000&j3\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*j5\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
  environments: <
    key: "zgRQRyxwxDbeam:env:docker:v1"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:timer:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:custom_window:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:param_windowed_value:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:sharded_key:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:harness_monitoring_infos:v1"
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:protocol:state_caching:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.37.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nL/tmp/artifacts/icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar\022@79bbed34591f15783869a605b839f02a9c034a3cec699aa5aafe622b13d3afe5"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n=icedtea-sound-ebvtNFkfFXg4aaYFuDnwKpwDSjzsaZqlqv5iKxPTr-U.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar\022@50b15308fb1be9c2d86747f5046d4541f73398735a642c7ca655ee4432a906a1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7jaccess-ULFTCPsb6cLYZ0f1BG1FQfczmHNaZCx8plXuRDKpBqE.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nI/tmp/artifacts/localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar\022@7541eac86c5a4d50a37d323c31c90f61aad9dff9b07fadae764c5a1e2d5a2a7b"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n:localedata-dUHqyGxaTVCjfTI8MckPYarZ3_mwf62udkxaHi1aKns.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar\022@5dd5277501aba1738ff4d0ac7c84e904445871b6c65d51cb8dbbcd5970a1fb70"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7nashorn-XdUndQGroXOP9NCsfITpBERYcbbGXVHLjbvNWXCh-3A.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nG/tmp/artifacts/cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar\022@62acee297d509cb08ea347308ca45d0611ab8a9fe5b4825983e00f4fad2d50f0"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n8cldrdata-YqzuKX1QnLCOo0cwjKRdBhGrip_ltIJZg-APT60tUPA.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nD/tmp/artifacts/dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar\022@761129d7ceae74417a5fa72166eb3e449cd6466ce571cc71d7f6ed95759ac952"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-dhEp186udEF6X6chZus-RJzWRmzlccxx1_btlXWayVI.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-kTA9K1krhOJ9gdisYu_EQwLhuMbR6lpQm_JapiyjCrw.jar\022@91303d2b592b84e27d81d8ac62efc44302e1b8c6d1ea5a509bf25aa62ca30abc"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\ncbeam-sdks-java-io-expansion-service-2.37.0-SNAPSHOT-kTA9K1krhOJ9gdisYu_EQwLhuMbR6lpQm_JapiyjCrw.jar"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2022/01/31 12:39:44 Prepared job with id: go-job-2-1643632784349239404_af7e7e62-9fed-4500-b22a-ef0eb3b98223 and staging token: go-job-2-1643632784349239404_af7e7e62-9fed-4500-b22a-ef0eb3b98223
2022/01/31 12:39:44 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1643632784359897705
2022/01/31 12:39:45 Staged binary artifact with token: 
2022/01/31 12:39:45 Submitted job: go0job0201643632784349239404-jenkins-0131123945-139cd1b4_b7b12b23-de3e-411e-9aa2-d7f0d0094855
2022/01/31 12:39:45 Job state: STOPPED
2022/01/31 12:39:45 Job state: STARTING
2022/01/31 12:39:45 Job state: RUNNING
2022/01/31 12:40:11 Job state: DONE
2022/01/31 12:40:11 Warning: 6 errors during metrics processing: [failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"zgRQRyxwxDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"zgRQRyxwxDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xbcA\xbcA\xbcA"  labels:{key:"PCOLLECTION"  value:"zgRQRyxwxDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"zgRQRyxwxDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"  type:"beam:metrics:sum_int64:v1"  payload:"\x01"  labels:{key:"PCOLLECTION"  value:"zgRQRyxwxDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1"  type:"beam:metrics:distribution_int64:v1"  payload:"\x01\xc9A\xc9A\xc9A"  labels:{key:"PCOLLECTION"  value:"zgRQRyxwxDExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}]
--- PASS: TestKafkaIO_BasicReadWrite (42.91s)
PASS
ok  	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	45.936s
FAIL

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingPython
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 36m 57s
217 actionable tasks: 152 executed, 56 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/chxueva5q73p2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark3 #1241

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/1241/display/redirect>

Changes:


------------------------------------------
[...truncated 326.58 KB...]
self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_tagged_join>

    def test_tagged_join(self):
      with TestPipeline() as p:
        enrich = (
            p | "Create enrich" >> beam.Create(
                [Enrich(1, "a"), Enrich(2, "b"), Enrich(26, "z")]))
        simple = (
            p | "Create simple" >> beam.Create([
                SimpleRow(1, "foo", 3.14),
                SimpleRow(26, "bar", 1.11),
                SimpleRow(1, "baz", 2.34)
            ]))
        out = ({
            'simple': simple, 'enrich': enrich
        }
               | SqlTransform(
                   """
                SELECT
                  simple.`id` AS `id`,
                  enrich.metadata AS metadata
                FROM simple
                JOIN enrich
                ON simple.`id` = enrich.`id`"""))
>       assert_that(out, equal_to([(1, "a"), (26, "z"), (1, "a")]))

apache_beam/transforms/sql_test.py:149: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7fba4c0482b0>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
__________________ SqlTransformTest.test_windowing_before_sql __________________

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_windowing_before_sql>

    def test_windowing_before_sql(self):
      with TestPipeline() as p:
        out = (
            p | beam.Create([
                SimpleRow(5, "foo", 1.),
                SimpleRow(15, "bar", 2.),
                SimpleRow(25, "baz", 3.)
            ])
            | beam.Map(lambda v: beam.window.TimestampedValue(v, v.id)).
            with_output_types(SimpleRow)
            | beam.WindowInto(
                beam.window.FixedWindows(10)).with_output_types(SimpleRow)
            | SqlTransform("SELECT COUNT(*) as `count` FROM PCOLLECTION"))
>       assert_that(out, equal_to([(1, ), (1, ), (1, )]))

apache_beam/transforms/sql_test.py:183: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7fba51f2fc50>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
_________________ SqlTransformTest.test_zetasql_generate_data __________________

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_zetasql_generate_data>

    def test_zetasql_generate_data(self):
      with TestPipeline() as p:
        out = p | SqlTransform(
            """SELECT
              CAST(1 AS INT64) AS `int`,
              CAST('foo' AS STRING) AS `str`,
              CAST(3.14  AS FLOAT64) AS `flt`""",
            dialect="zetasql")
>       assert_that(out, equal_to([(1, "foo", 3.14)]))

apache_beam/transforms/sql_test.py:168: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:596: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:573: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/portability/portable_runner.py:438: in run_pipeline
    job_service_handle = self.create_job_service(options)
apache_beam/runners/portability/portable_runner.py:317: in create_job_service
    return self.create_job_service_handle(server.start(), options)
apache_beam/runners/portability/job_server.py:54: in start
    grpc.channel_ready_future(channel).result(timeout=self._timeout)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:139: in result
    self._block(timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <grpc._utilities._ChannelReadyFuture object at 0x7fba4c2279e8>
timeout = 60

    def _block(self, timeout):
        until = None if timeout is None else time.time() + timeout
        with self._condition:
            while True:
                if self._cancelled:
                    raise grpc.FutureCancelledError()
                elif self._matured:
                    return
                else:
                    if until is None:
                        self._condition.wait()
                    else:
                        remaining = until - time.time()
                        if remaining < 0:
>                           raise grpc.FutureTimeoutError()
E                           grpc.FutureTimeoutError

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_utilities.py:85: FutureTimeoutError
----------------------------- Captured stderr call -----------------------------
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
------------------------------ Captured log call -------------------------------
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/python/pytest_xlangSqlValidateRunner.xml> -
================= 9 failed, 5167 deselected in 659.77 seconds ==================

> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:3:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:3:job-server:sparkJobServerCleanup

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark3/ws/src/sdks/go/test/build.gradle'> line: 170

* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:3:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 43m 24s
217 actionable tasks: 150 executed, 58 from cache, 9 up-to-date

Publishing build scan...
https://gradle.com/s/hsydlpx7bdhzy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org