You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/12/19 06:17:12 UTC

Build failed in Jenkins: beam_PostCommit_XVR_Spark #3143

See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/3143/display/redirect>

Changes:


------------------------------------------
[...truncated 441.96 KB...]
  Created wheel for apache-beam: filename=apache_beam-2.36.0.dev0-cp36-cp36m-linux_x86_64.whl size=11013342 sha256=6dda672e55ef44ab1d508f6a36e9a139a2fdd9463acf80395d84ada4e32ac76e
  Stored in directory: /root/.cache/pip/wheels/71/0a/9d/8434df5e361fe8c44220318cb21455c5dcaa3d25589894f9c8
Successfully built apache-beam
Installing collected packages: apache-beam
Successfully installed apache-beam-2.36.0.dev0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
WARNING: You are using pip version 21.2.4; however, version 21.3.1 is available.
You should consider upgrading via the '/usr/local/bin/python -m pip install --upgrade pip' command.
Removing intermediate container 3af3594bbc00
 ---> 0514c6ec673e
Step 14/31 : RUN pip check || (echo "Container does not include required Beam dependencies or has conflicting dependencies. If Beam dependencies have changed, you need to regenerate base_image_requirements.txt files. See: https://s.apache.org/beam-python-requirements-generate" && exit 1)
 ---> Running in 4d97308c14a5
No broken requirements found.
Removing intermediate container 4d97308c14a5
 ---> 67202b3acc62
Step 15/31 : RUN pip freeze --all
 ---> Running in 0569d0dac5bf
absl-py==0.15.0
apache-beam @ file:///opt/apache/beam/tars/apache-beam.tar.gz
astunparse==1.6.3
atomicwrites==1.4.0
attrs==21.2.0
beautifulsoup4==4.10.0
bs4==0.0.1
cached-property==1.5.2
cachetools==4.2.4
certifi==2021.10.8
charset-normalizer==2.0.8
clang==5.0
click==8.0.3
cloudpickle==2.0.0
crcmod==1.7
Cython==0.29.24
dataclasses==0.8
deprecation==2.1.0
dill==0.3.1.1
docker==5.0.3
docopt==0.6.2
execnet==1.9.0
fastavro==1.4.7
fasteners==0.16.3
flatbuffers==1.12
freezegun==1.1.0
future==0.18.2
gast==0.4.0
google-api-core==1.31.4
google-api-python-client==2.31.0
google-apitools==0.5.31
google-auth==1.35.0
google-auth-httplib2==0.1.0
google-auth-oauthlib==0.4.6
google-cloud-bigquery==2.30.1
google-cloud-bigquery-storage==2.10.1
google-cloud-bigtable==1.7.0
google-cloud-core==1.7.2
google-cloud-datastore==1.15.3
google-cloud-dlp==3.3.1
google-cloud-language==1.3.0
google-cloud-profiler==3.0.6
google-cloud-pubsub==1.7.0
google-cloud-recommendations-ai==0.2.0
google-cloud-spanner==1.19.1
google-cloud-videointelligence==1.16.1
google-cloud-vision==1.0.0
google-crc32c==1.3.0
google-pasta==0.2.0
google-python-cloud-debugger==2.18
google-resumable-media==2.1.0
googleapis-common-protos==1.53.0
greenlet==1.1.2
grpc-google-iam-v1==0.12.3
grpcio==1.42.0
grpcio-gcp==0.2.2
guppy3==3.1.2
h5py==3.1.0
hdfs==2.6.0
httplib2==0.19.1
idna==3.3
importlib-metadata==4.8.2
joblib==1.1.0
keras==2.6.0
Keras-Preprocessing==1.1.2
libcst==0.3.23
Markdown==3.3.6
mmh3==3.0.0
mock==2.0.0
more-itertools==8.12.0
mypy-extensions==0.4.3
nltk==3.6.5
nose==1.3.7
numpy==1.19.5
oauth2client==4.1.3
oauthlib==3.1.1
opt-einsum==3.3.0
orjson==3.6.1
packaging==21.3
pandas==1.1.5
parameterized==0.7.5
pbr==5.8.0
pip==21.2.4
pluggy==0.13.1
proto-plus==1.19.8
protobuf==3.19.1
psycopg2-binary==2.9.2
py==1.11.0
pyarrow==6.0.1
pyasn1==0.4.8
pyasn1-modules==0.2.8
pydot==1.4.2
PyHamcrest==1.10.1
pymongo==3.12.2
pyparsing==2.4.7
pytest==4.6.11
pytest-forked==1.3.0
pytest-timeout==1.4.2
pytest-xdist==1.34.0
python-dateutil==2.8.2
python-snappy==0.6.0
pytz==2021.3
PyYAML==6.0
regex==2021.11.10
requests==2.26.0
requests-mock==1.9.3
requests-oauthlib==1.3.0
rsa==4.8
scikit-learn==0.24.2
scipy==1.5.4
setuptools==57.5.0
six==1.15.0
soupsieve==2.3.1
SQLAlchemy==1.4.27
tenacity==5.1.5
tensorboard==2.6.0
tensorboard-data-server==0.6.1
tensorboard-plugin-wit==1.8.0
tensorflow==2.6.2
tensorflow-estimator==2.6.0
termcolor==1.1.0
testcontainers==3.4.2
threadpoolctl==3.0.0
tqdm==4.62.3
typing-extensions==3.7.4.3
typing-inspect==0.7.1
uritemplate==4.1.1
urllib3==1.26.7
wcwidth==0.2.5
websocket-client==1.2.1
Werkzeug==2.0.2
wheel==0.37.0
wrapt==1.12.1
zipp==3.6.0
Removing intermediate container 0569d0dac5bf
 ---> d9eecfea22f2
Step 16/31 : COPY target/LICENSE /opt/apache/beam/
 ---> bc105d66170c
Step 17/31 : COPY target/LICENSE.python /opt/apache/beam/
 ---> ee433f66e301
Step 18/31 : COPY target/NOTICE /opt/apache/beam/
 ---> a8cd46a46e76
Step 19/31 : COPY target/launcher/linux_amd64/boot /opt/apache/beam/
 ---> a38c13b6e73d
Step 20/31 : ENTRYPOINT ["/opt/apache/beam/boot"]
 ---> Running in e4fac712e4e2
Removing intermediate container e4fac712e4e2
 ---> 800a129c5235
Step 21/31 : FROM beam as third_party_licenses
 ---> 800a129c5235
Step 22/31 : ARG pull_licenses
 ---> Running in e9b79f4485dc
Removing intermediate container e9b79f4485dc
 ---> b4a72270c014
Step 23/31 : COPY target/license_scripts /tmp/license_scripts/
 ---> 7c0b2c2a02f7
Step 24/31 : COPY target/LICENSE target/go-licenses/* /opt/apache/beam/third_party_licenses/golang/
 ---> 833b59e37096
Step 25/31 : RUN rm /opt/apache/beam/third_party_licenses/golang/LICENSE
 ---> Running in d3ff5b6b97f2
Removing intermediate container d3ff5b6b97f2
 ---> 6c7f2866b6c6
Step 26/31 : COPY target/license_scripts /tmp/license_scripts/
 ---> 39593e70214f
Step 27/31 : RUN if [ "$pull_licenses" = "true" ] ; then       pip install 'pip-licenses<3.0.0' pyyaml tenacity &&       python /tmp/license_scripts/pull_licenses_py.py ;     fi
 ---> Running in 71ec563e31f4
Collecting pip-licenses<3.0.0
  Downloading pip_licenses-2.3.0-py3-none-any.whl (14 kB)
Requirement already satisfied: pyyaml in /usr/local/lib/python3.6/site-packages (6.0)
Requirement already satisfied: tenacity in /usr/local/lib/python3.6/site-packages (5.1.5)
Collecting PTable
  Downloading PTable-0.9.2.tar.gz (31 kB)
Requirement already satisfied: six>=1.9.0 in /usr/local/lib/python3.6/site-packages (from tenacity) (1.15.0)
Building wheels for collected packages: PTable
  Building wheel for PTable (setup.py): started
  Building wheel for PTable (setup.py): finished with status 'done'
  Created wheel for PTable: filename=PTable-0.9.2-py3-none-any.whl size=22924 sha256=bbb2008054d2f19e4a96f0447b4283277476db04ebf7ee085107cde0393fd86f
  Stored in directory: /root/.cache/pip/wheels/f3/65/67/71f473ec87ea4554d1bdfb1b5128cfe1414c2b113b72a1ee8e
Successfully built PTable
Installing collected packages: PTable, pip-licenses
Successfully installed PTable-0.9.2 pip-licenses-2.3.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
WARNING: You are using pip version 21.2.4; however, version 21.3.1 is available.
You should consider upgrading via the '/usr/local/bin/python -m pip install --upgrade pip' command.
INFO:root:Successfully pulled licenses for 130 dependencies
Skip pulling license for  bs4
Removing intermediate container 71ec563e31f4
 ---> 9a9251a56698
Step 28/31 : FROM beam
 ---> 800a129c5235
Step 29/31 : ARG pull_licenses
 ---> Running in d94a80ee7487
Removing intermediate container d94a80ee7487
 ---> f24f361aed50
Step 30/31 : COPY --from=third_party_licenses /opt/apache/beam/third_party_licenses /opt/apache/beam/third_party_licenses
 ---> c90756e1750c
Step 31/31 : RUN if [ "$pull_licenses" != "true" ] ; then       rm -rf /opt/apache/beam/third_party_licenses ;     fi
 ---> Running in 2e72e777e6f3
Removing intermediate container 2e72e777e6f3
 ---> 2da6db60c9df
Successfully built 2da6db60c9df
Successfully tagged apache/beam_python3.6_sdk:2.36.0.dev

> Task :runners:spark:2:job-server:sparkJobServerCleanup

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/java/io/synthetic/build.gradle'> line: 20

* What went wrong:
A problem occurred evaluating project ':sdks:java:io:synthetic'.
> Could not create task ':sdks:java:io:synthetic:projects'.
   > java.util.ConcurrentModificationException (no error message)

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 14m 34s
200 actionable tasks: 140 executed, 53 from cache, 7 up-to-date

Publishing build scan...
https://gradle.com/s/iclpmfsrzyalu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_XVR_Spark #3147

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/3147/display/redirect>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark #3146

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/3146/display/redirect>

Changes:


------------------------------------------
[...truncated 642.69 KB...]
    >
  >
  coders: <
    key: "pLAMlDxKyQKvCoder1"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "pLAMlDxKyQSerializableCoder"
      component_coder_ids: "pLAMlDxKyQVoidCoder"
    >
  >
  coders: <
    key: "pLAMlDxKyQKvCoder2"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "pLAMlDxKyQVarIntCoder"
      component_coder_ids: "pLAMlDxKyQSerializableCoder1"
    >
  >
  coders: <
    key: "pLAMlDxKyQKvCoder3"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "pLAMlDxKyQVarIntCoder"
      component_coder_ids: "pLAMlDxKyQTimestampedValueCoder"
    >
  >
  coders: <
    key: "pLAMlDxKyQKvCoder4"
    value: <
      spec: <
        urn: "beam:coder:kv:v1"
      >
      component_coder_ids: "pLAMlDxKyQVarIntCoder"
      component_coder_ids: "pLAMlDxKyQIterableCoder"
    >
  >
  coders: <
    key: "pLAMlDxKyQSerializableCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\241\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000$N1\000\\io.BoundedSource!`v\023\177\370QA\0015\010r\000\035Z3\000\t,0\020\213\372\020Y\214\347\323\002\000\000xp"
      >
    >
  >
  coders: <
    key: "pLAMlDxKyQSerializableCoder1"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\312\214\002\360^\254\355\000\005sr\000,org.apache.beam.sdk.coders.SerializableCoder\010\364 \216\305ynl\002\000\001L\000\004typet\000\021Ljava/lang/Class;xr\000&ofV\000\024Custom\005P8j\260\010\235\013;\035\013\002\000\000xr\000 j\213\000\005/\034C\335\325\211\256\274~\370\001/\020pvr\000;N1\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "pLAMlDxKyQTimestampedValueCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001?\315\003\270\254\355\000\005sr\000Aorg.apache.beam.sdk.values.TimestampedV\001\022\000$>\021\000LCoder\364\376\202p\274iX-\002\000\001L\000\nv\001)\005\030\034t\000\"Lorg/\t]\000/\001],/sdk/coders/\005$\020;xr\000*N\202\000\t%(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000,j1\000(Serializabl\t\277\034\010\364 \216\305ynl\005\327\\\004typet\000\021Ljava/lang/Class\001\300\000&jV\000\024Custom\005\215Pj\260\010\235\013;\035\013\002\000\000xq\000~\000\003vr\000;N;\000Pio.BoundedReadFromUnb\t\021`Source$Shardd\236\240U\205\000j\363\002\000\000xp"
      >
    >
  >
  coders: <
    key: "pLAMlDxKyQValueWithRecordIdCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\001h\364\004\320\254\355\000\005sr\000Corg.apache.beam.sdk.values.ValueWithRecordId$B\022\000LCoder{\343\376bQ\014\204>\002\000\002L\000\007i\t\025\034t\000+Lorg/\t\\\000/\001\\\024/sdk/c\0012(s/ByteArray\005B\014;L\000\n\005u\005\016\010t\000\"n;\000\005$\020;xr\000*N\274\000\t`(.Structured\005/8s\277\022\016\325\3246\021\002\000\000xr\000 j9\000\005/\034C\335\325\211\256\274~\370\001/\020psr\000)j1\0006\312\000\034\351\277\005|\263\3746\333\001:\010r\000&j8\000\024Atomic\005o\034\307\354\265\314\205tPF\0015 q\000~\000\003sr\000-N;\000$io.kafka.K\001\006\000R%\227\005B<I\320\251Y\032\341{#\002\000\001L\000\007kv\005\025\010t\000$nH\001\000K\t&\000;\025s\000\"Ns\000-P\r- j\000\275\271\035\243\3126\002!\353\010\010ke)\277\001\264\000\0022\303\001\005\022\rY$q\000~\000\006q\000~\000\016"
      >
    >
  >
  coders: <
    key: "pLAMlDxKyQVarIntCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\330\001\350\254\355\000\005sr\000&org.apache.beam.sdk.coders.VarIntCoder\300K2\332\251KVh\002\000\000xrr5\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*jj\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  coders: <
    key: "pLAMlDxKyQVoidCoder"
    value: <
      spec: <
        urn: "beam:coders:javasdk:0.1"
        payload: "\202SNAPPY\000\000\000\000\001\000\000\000\001\000\000\000\210\326\001\350\254\355\000\005sr\000$org.apache.beam.sdk.coders.VoidCoder\271\277U\233\350\r\257U\002\000\000xr\000&j3\000\024Atomic\0055 \307\354\265\314\205tPF\002\0055\000*j5\000$Structured\0059\034s\277\022\016\325\3246\021\t9\000 j9\000\005/0C\335\325\211\256\274~\370\002\000\000xp"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
  environments: <
    key: "pLAMlDxKyQbeam:env:docker:v1"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n apache/beam_java8_sdk:2.36.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:timer:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:custom_window:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:param_windowed_value:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:sharded_key:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:harness_monitoring_infos:v1"
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.36.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nL/tmp/artifacts/icedtea-sound-Bdoi2wYa757-fzq5vconCy4SSQ22ZaOq7yuC98fKPs8.jar\022@05da22db061aef9efe7f3ab9bdca270b2e12490db665a3aaef2b82f7c7ca3ecf"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n=icedtea-sound-Bdoi2wYa757-fzq5vconCy4SSQ22ZaOq7yuC98fKPs8.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/jaccess-CMbK-IOdQPLKHEqCuDnE-yBk-VpbtVT-hgjbHRUGO78.jar\022@08c6caf8839d40f2ca1c4a82b839c4fb2064f95a5bb554fe8608db1d15063bbf"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7jaccess-CMbK-IOdQPLKHEqCuDnE-yBk-VpbtVT-hgjbHRUGO78.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nI/tmp/artifacts/localedata-ae5Z0L6ak4922fztWeWy7ajiWXdG3ubNrwerJRFoFj0.jar\022@69ee59d0be9a938f76d9fced59e5b2eda8e2597746dee6cdaf07ab251168163d"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n:localedata-ae5Z0L6ak4922fztWeWy7ajiWXdG3ubNrwerJRFoFj0.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nF/tmp/artifacts/nashorn-XHtz_UehGpYcLTOrATrTnMNVUgEVa_ttoWkPxnVfqTo.jar\022@5c7b73fd47a11a961c2d33ab013ad39cc3555201156bfb6da1690fc6755fa93a"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n7nashorn-XHtz_UehGpYcLTOrATrTnMNVUgEVa_ttoWkPxnVfqTo.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nG/tmp/artifacts/cldrdata-k07I6K9W3X5KTQbcDIEsqM0LXyM18f0eR6IaJw-P_xk.jar\022@934ec8e8af56dd7e4a4d06dc0c812ca8cd0b5f2335f1fd1e47a21a270f8fff19"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n8cldrdata-k07I6K9W3X5KTQbcDIEsqM0LXyM18f0eR6IaJw-P_xk.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nD/tmp/artifacts/dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar\022@4468420e0dc654e4150b6afa91ad8dd21988e1ea901fa5686eea009d0ef83271"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.36.0-SNAPSHOT-NgyoWvk4sI87PKOgAm2tDaLwjvTUJ2Lm6vCacMiUhPY.jar\022@360ca85af938b08f3b3ca3a0026dad0da2f08ef4d42762e6eaf09a70c89484f6"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\ncbeam-sdks-java-io-expansion-service-2.36.0-SNAPSHOT-NgyoWvk4sI87PKOgAm2tDaLwjvTUJ2Lm6vCacMiUhPY.jar"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2021/12/20 00:15:15 Prepared job with id: go-job-2-1639959315355528346_5ab8930b-a317-4924-a208-ddca885bbfef and staging token: go-job-2-1639959315355528346_5ab8930b-a317-4924-a208-ddca885bbfef
2021/12/20 00:15:15 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1639959315366914228
2021/12/20 00:15:16 Staged binary artifact with token: 
2021/12/20 00:15:16 Submitted job: go0job0201639959315355528346-jenkins-1220001516-28502a59_383ac99f-6712-4798-82d6-1b96cf5ffffe
2021/12/20 00:15:16 Job state: STOPPED
2021/12/20 00:15:16 Job state: STARTING
2021/12/20 00:15:16 Job state: RUNNING
2021/12/20 00:15:40 Job state: DONE
2021/12/20 00:15:40 Warning: 6 errors during metrics processing: [failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"pLAMlDxKyQExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1" type:"beam:metrics:distribution_int64:v1" payload:"\x01\xc9A\xc9A\xc9A" labels:{key:"PCOLLECTION" value:"pLAMlDxKyQExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"pLAMlDxKyQExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"pLAMlDxKyQExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1" type:"beam:metrics:distribution_int64:v1" payload:"\x01\xc9A\xc9A\xc9A" labels:{key:"PCOLLECTION" value:"pLAMlDxKyQExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1" type:"beam:metrics:distribution_int64:v1" payload:"\x01\xbcA\xbcA\xbcA" labels:{key:"PCOLLECTION" value:"pLAMlDxKyQExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"}]
--- PASS: TestKafkaIO_BasicReadWrite (44.44s)
PASS
ok  	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	47.460s

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingJava
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:2:job-server:sparkJobServerCleanup

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 37m 5s
213 actionable tasks: 152 executed, 54 from cache, 7 up-to-date

Publishing build scan...
https://gradle.com/s/6afejhpgljrce

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark #3145

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/3145/display/redirect>

Changes:


------------------------------------------
[...truncated 692.01 KB...]
        else:
>           raise _InactiveRpcError(state)
E           grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E           	status = StatusCode.UNAVAILABLE
E           	details = "failed to connect to all addresses"
E           	debug_error_string = "{"created":"@1639938200.410500412","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3134,"referenced_errors":[{"created":"@1639938200.410499405","description":"failed to connect to all addresses","file":"src/core/lib/transport/error_utils.cc","file_line":163,"grpc_status":14}]}"
E           >

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError
________ ValidateRunnerXlangTest.test_multi_input_output_with_sideinput ________

self = <apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest testMethod=test_multi_input_output_with_sideinput>
test_pipeline = None

    @pytest.mark.xlang_transforms
    def test_multi_input_output_with_sideinput(self, test_pipeline=None):
      CrossLanguageTestPipelines().run_multi_input_output_with_sideinput(
>         test_pipeline or self.create_pipeline())

apache_beam/transforms/validate_runner_xlang_test.py:259: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/transforms/validate_runner_xlang_test.py:119: in run_multi_input_output_with_sideinput
    TEST_MULTI_URN, None, self.expansion_service)
apache_beam/transforms/ptransform.py:614: in __ror__
    result = p.apply(self, pvalueish, label)
apache_beam/pipeline.py:708: in apply
    pvalueish_result = self.runner.apply(transform, pvalueish, self._options)
apache_beam/runners/runner.py:185: in apply
    return m(transform, input, options)
apache_beam/runners/runner.py:215: in apply_PTransform
    return transform.expand(input)
apache_beam/transforms/external.py:501: in expand
    response = service.Expand(request)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:946: in __call__
    return _end_unary_response_blocking(state, call, False, None)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

state = <grpc._channel._RPCState object at 0x7f1213a86630>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f12139d4fc8>
with_call = False, deadline = None

    def _end_unary_response_blocking(state, call, with_call, deadline):
        if state.code is grpc.StatusCode.OK:
            if with_call:
                rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
                return state.response, rendezvous
            else:
                return state.response
        else:
>           raise _InactiveRpcError(state)
E           grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E           	status = StatusCode.UNAVAILABLE
E           	details = "failed to connect to all addresses"
E           	debug_error_string = "{"created":"@1639938200.566415220","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3134,"referenced_errors":[{"created":"@1639938200.566414259","description":"failed to connect to all addresses","file":"src/core/lib/transport/error_utils.cc","file_line":163,"grpc_status":14}]}"
E           >

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError
____________________ ValidateRunnerXlangTest.test_partition ____________________

self = <apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest testMethod=test_partition>
test_pipeline = None

    @pytest.mark.xlang_transforms
    def test_partition(self, test_pipeline=None):
      CrossLanguageTestPipelines().run_partition(
>         test_pipeline or self.create_pipeline())

apache_beam/transforms/validate_runner_xlang_test.py:290: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/transforms/validate_runner_xlang_test.py:235: in run_partition
    TEST_PARTITION_URN, None, self.expansion_service))
apache_beam/pvalue.py:137: in __or__
    return self.pipeline.apply(ptransform, self)
apache_beam/pipeline.py:708: in apply
    pvalueish_result = self.runner.apply(transform, pvalueish, self._options)
apache_beam/runners/runner.py:185: in apply
    return m(transform, input, options)
apache_beam/runners/runner.py:215: in apply_PTransform
    return transform.expand(input)
apache_beam/transforms/external.py:501: in expand
    response = service.Expand(request)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:946: in __call__
    return _end_unary_response_blocking(state, call, False, None)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

state = <grpc._channel._RPCState object at 0x7f1213a50630>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f1213a602c8>
with_call = False, deadline = None

    def _end_unary_response_blocking(state, call, with_call, deadline):
        if state.code is grpc.StatusCode.OK:
            if with_call:
                rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
                return state.response, rendezvous
            else:
                return state.response
        else:
>           raise _InactiveRpcError(state)
E           grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E           	status = StatusCode.UNAVAILABLE
E           	details = "failed to connect to all addresses"
E           	debug_error_string = "{"created":"@1639938200.725163482","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3134,"referenced_errors":[{"created":"@1639938200.725162494","description":"failed to connect to all addresses","file":"src/core/lib/transport/error_utils.cc","file_line":163,"grpc_status":14}]}"
E           >

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError
_____________________ ValidateRunnerXlangTest.test_prefix ______________________

self = <apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest testMethod=test_prefix>
test_pipeline = None

    @pytest.mark.xlang_transforms
    def test_prefix(self, test_pipeline=None):
      CrossLanguageTestPipelines().run_prefix(
>         test_pipeline or self.create_pipeline())

apache_beam/transforms/validate_runner_xlang_test.py:254: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/transforms/validate_runner_xlang_test.py:98: in run_prefix
    self.expansion_service))
apache_beam/pvalue.py:137: in __or__
    return self.pipeline.apply(ptransform, self)
apache_beam/pipeline.py:708: in apply
    pvalueish_result = self.runner.apply(transform, pvalueish, self._options)
apache_beam/runners/runner.py:185: in apply
    return m(transform, input, options)
apache_beam/runners/runner.py:215: in apply_PTransform
    return transform.expand(input)
apache_beam/transforms/external.py:501: in expand
    response = service.Expand(request)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:946: in __call__
    return _end_unary_response_blocking(state, call, False, None)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

state = <grpc._channel._RPCState object at 0x7f1213c786a0>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f1213a4fd08>
with_call = False, deadline = None

    def _end_unary_response_blocking(state, call, with_call, deadline):
        if state.code is grpc.StatusCode.OK:
            if with_call:
                rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
                return state.response, rendezvous
            else:
                return state.response
        else:
>           raise _InactiveRpcError(state)
E           grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E           	status = StatusCode.UNAVAILABLE
E           	details = "failed to connect to all addresses"
E           	debug_error_string = "{"created":"@1639938201.035132961","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3134,"referenced_errors":[{"created":"@1639938201.035131986","description":"failed to connect to all addresses","file":"src/core/lib/transport/error_utils.cc","file_line":163,"grpc_status":14}]}"
E           >

../../build/gradleenv/1922375555/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/pytest_xlangValidateRunner.xml> -
============ 8 failed, 2 skipped, 5131 deselected in 10.77 seconds =============

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython FAILED

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql
>>> RUNNING integration tests with pipeline options: --runner=PortableRunner --job_endpoint=localhost:36911 --environment_cache_millis=10000 --experiments=beam_fn_api
>>>   pytest options: 
>>>   collect markers: -m=xlang_sql_expansion_service
============================= test session starts ==============================
platform linux -- Python 3.6.13, pytest-4.6.11, py-1.11.0, pluggy-0.13.1
rootdir: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python,> inifile: pytest.ini
plugins: xdist-1.34.0, timeout-1.4.2, forked-1.4.0, requests-mock-1.9.3
timeout: 600.0s
timeout method: signal
timeout func_only: False
collected 5141 items / 5132 deselected / 9 selected

apache_beam/transforms/sql_test.py .........                             [100%]

- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/pytest_xlangSqlValidateRunner.xml> -
================= 9 passed, 5132 deselected in 181.25 seconds ==================

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerCleanup
Stopping expansion service pid: 28745.
Stopping expansion service pid: 28746.

> Task :runners:spark:2:job-server:sparkJobServerCleanup
Stopping job server pid: 10831.

FAILURE: Build completed with 5 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingJava'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingJava/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingPython/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingPythonOnly/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

5: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 24m 48s
213 actionable tasks: 152 executed, 54 from cache, 7 up-to-date

Publishing build scan...
https://gradle.com/s/3jeenjbrjtpk4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_XVR_Spark #3144

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/3144/display/redirect>

Changes:


------------------------------------------
[...truncated 647.49 KB...]
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: "\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.36.0-SNAPSHOT-NgyoWvk4sI87PKOgAm2tDaLwjvTUJ2Lm6vCacMiUhPY.jar\022@360ca85af938b08f3b3ca3a0026dad0da2f08ef4d42762e6eaf09a70c89484f6"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\ncbeam-sdks-java-io-expansion-service-2.36.0-SNAPSHOT-NgyoWvk4sI87PKOgAm2tDaLwjvTUJ2Lm6vCacMiUhPY.jar"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v0"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:version:sdk_base:go"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      dependencies: <
        type_urn: "beam:artifact:type:go_worker_binary:v1"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n\006worker"
      >
    >
  >
>
root_transform_ids: "s1"
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2021/12/19 12:17:30 Prepared job with id: go-job-2-1639916250659026207_149702ee-56e2-4bcc-b0bc-5c28d3608672 and staging token: go-job-2-1639916250659026207_149702ee-56e2-4bcc-b0bc-5c28d3608672
2021/12/19 12:17:30 Cross-compiling <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go> as /tmp/worker-2-1639916250690805347
2021/12/19 12:17:33 Staged binary artifact with token: 
2021/12/19 12:17:33 Submitted job: go0job0201639916250659026207-jenkins-1219121733-67cdf5e2_98322cef-1a7f-41ea-97cc-e627f2938d80
2021/12/19 12:17:33 Job state: STOPPED
2021/12/19 12:17:33 Job state: STARTING
2021/12/19 12:17:33 Job state: RUNNING
2021/12/19 12:18:08 Job state: DONE
2021/12/19 12:18:08 Warning: 6 errors during metrics processing: [failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"FtIwrNzbVGExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1" type:"beam:metrics:distribution_int64:v1" payload:"\x01\xc9A\xc9A\xc9A" labels:{key:"PCOLLECTION" value:"FtIwrNzbVGExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"FtIwrNzbVGExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1" type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION" value:"FtIwrNzbVGExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1" type:"beam:metrics:distribution_int64:v1" payload:"\x01\xbcA\xbcA\xbcA" labels:{key:"PCOLLECTION" value:"FtIwrNzbVGExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"} failed to deduce Step from MonitoringInfo: urn:"beam:metric:sampled_byte_size:v1" type:"beam:metrics:distribution_int64:v1" payload:"\x01\xc9A\xc9A\xc9A" labels:{key:"PCOLLECTION" value:"FtIwrNzbVGExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}]
--- PASS: TestKafkaIO_BasicReadWrite (64.06s)
PASS
ok  	github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka	67.110s

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingJava

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > combineGloballyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:257
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > partitionTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:314
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > coGroupByKeyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:230
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > groupByKeyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:198
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > combinePerKeyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:275
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > flattenTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:295
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > singleInputOutputTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:156
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > multiInputOutputWithSideInputTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:178
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

8 tests completed, 8 failed

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingJava FAILED

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > combineGloballyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:257
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > partitionTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:314
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > coGroupByKeyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:230
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > groupByKeyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:198
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > combinePerKeyTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:275
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > flattenTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:295
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > singleInputOutputTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:156
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > multiInputOutputWithSideInputTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:178
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

8 tests completed, 8 failed

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython FAILED

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly

org.apache.beam.runners.core.construction.ValidateRunnerXlangTest > pythonDependenciesTest FAILED
    org.apache.beam.vendor.grpc.v1p36p0.io.grpc.StatusRuntimeException at ValidateRunnerXlangTest.java:338
        Caused by: org.apache.beam.vendor.grpc.v1p36p0.io.netty.channel.AbstractChannel$AnnotatedConnectException
            Caused by: java.net.ConnectException at Errors.java:124
    java.lang.NullPointerException at ValidateRunnerXlangTest.java:123

1 test completed, 1 failed

> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:2:job-server:sparkJobServerCleanup

FAILURE: Build completed with 5 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingJava'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingJava/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingPython/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPythonOnly'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingPythonOnly/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

4: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

5: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingPython'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.9.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 24m 38s
213 actionable tasks: 153 executed, 53 from cache, 7 up-to-date

Publishing build scan...
https://gradle.com/s/lezasb4tyfktc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org