You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/05/15 23:28:51 UTC

Build failed in Jenkins: beam_PostCommit_Python_ValidatesRunner_Dataflow #1654

See <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1654/display/redirect>

------------------------------------------
GitHub pull request #5372 of commit d79a0a4adf9d275d5db3312df405da99dbf5d5eb, no merge conflicts.
Setting status of d79a0a4adf9d275d5db3312df405da99dbf5d5eb to PENDING with url https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1654/ and message: 'Build started sha1 is merged.'
Using context: Jenkins: Google Cloud Dataflow Runner Python ValidatesRunner Tests
[EnvInject] - Loading node environment variables.
Building remotely on beam10 (beam) in workspace <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/>
 > git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/5372/*:refs/remotes/origin/pr/5372/*
 > git rev-parse refs/remotes/origin/pr/5372/merge^{commit} # timeout=10
 > git rev-parse refs/remotes/origin/origin/pr/5372/merge^{commit} # timeout=10
Checking out Revision c27d5fe0835562ae878b882c9e28e45aaf0d9315 (refs/remotes/origin/pr/5372/merge)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f c27d5fe0835562ae878b882c9e28e45aaf0d9315
Commit message: "Merge d79a0a4adf9d275d5db3312df405da99dbf5d5eb into 72b6c8f6d0923da641b5344eb6a1fa8c260d596c"
First time build. Skipping changelog.
Cleaning workspace
 > git rev-parse --verify HEAD # timeout=10
Resetting working tree
 > git reset --hard # timeout=10
 > git clean -fdx # timeout=10
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/gradlew> --info --continue --max-workers=16 -Dorg.gradle.jvmargs=-Xmx3328m :beam-sdks-python:validatesRunnerTests
Initialized native services in: /home/jenkins/.gradle/native
To honour the JVM settings for this build a new JVM will be forked. Please consider using the daemon: https://docs.gradle.org/4.7/userguide/gradle_daemon.html.
Starting process 'Gradle build daemon'. Working directory: /home/jenkins/.gradle/daemon/4.7 Command: /usr/local/asfpackages/java/jdk1.8.0_152/bin/java -Xmx3328m -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -cp /home/jenkins/.gradle/wrapper/dists/gradle-4.7-bin/73xux3q4721jgr4pjvqf9o30/gradle-4.7/lib/gradle-launcher-4.7.jar org.gradle.launcher.daemon.bootstrap.GradleDaemon 4.7
Successfully started process 'Gradle build daemon'
An attempt to start the daemon took 0.755 secs.
The client will now receive all logging from the daemon (pid: 19354). The daemon log file: /home/jenkins/.gradle/daemon/4.7/daemon-19354.out.log
Closing daemon's stdin at end of input.
The daemon will no longer process any standard input.
Daemon will be stopped at the end of the build stopping after processing
Using 16 worker leases.
Starting Build
Settings evaluated using settings file '<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/settings.gradle'.>
Using local directory build cache for the root build (location = /home/jenkins/.gradle/caches/build-cache-1, removeUnusedEntriesAfter = 7 days).
Projects loaded. Root project using build file '<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/build.gradle'.>
Included projects: [root project 'beam', project ':beam-examples-java', project ':beam-local-artifact-service-java', project ':beam-model-fn-execution', project ':beam-model-job-management', project ':beam-model-pipeline', project ':beam-runners-apex', project ':beam-runners-core-construction-java', project ':beam-runners-core-java', project ':beam-runners-direct-java', project ':beam-runners-extensions-java-metrics', project ':beam-runners-flink_2.11', project ':beam-runners-gcp-gcemd', project ':beam-runners-gcp-gcsproxy', project ':beam-runners-gearpump', project ':beam-runners-google-cloud-dataflow-java', project ':beam-runners-java-fn-execution', project ':beam-runners-local-java-core', project ':beam-runners-reference-java', project ':beam-runners-reference-job-server', project ':beam-runners-spark', project ':beam-sdks-go', project ':beam-sdks-go-container', project ':beam-sdks-go-examples', project ':beam-sdks-java-build-tools', project ':beam-sdks-java-container', project ':beam-sdks-java-core', project ':beam-sdks-java-extensions-google-cloud-platform-core', project ':beam-sdks-java-extensions-join-library', project ':beam-sdks-java-extensions-json-jackson', project ':beam-sdks-java-extensions-protobuf', project ':beam-sdks-java-extensions-sketching', project ':beam-sdks-java-extensions-sorter', project ':beam-sdks-java-extensions-sql', project ':beam-sdks-java-fn-execution', project ':beam-sdks-java-harness', project ':beam-sdks-java-io-amazon-web-services', project ':beam-sdks-java-io-amqp', project ':beam-sdks-java-io-cassandra', project ':beam-sdks-java-io-common', project ':beam-sdks-java-io-elasticsearch', project ':beam-sdks-java-io-elasticsearch-tests-2', project ':beam-sdks-java-io-elasticsearch-tests-5', project ':beam-sdks-java-io-elasticsearch-tests-common', project ':beam-sdks-java-io-file-based-io-tests', project ':beam-sdks-java-io-google-cloud-platform', project ':beam-sdks-java-io-hadoop-common', project ':beam-sdks-java-io-hadoop-file-system', project ':beam-sdks-java-io-hadoop-input-format', project ':beam-sdks-java-io-hbase', project ':beam-sdks-java-io-hcatalog', project ':beam-sdks-java-io-jdbc', project ':beam-sdks-java-io-jms', project ':beam-sdks-java-io-kafka', project ':beam-sdks-java-io-kinesis', project ':beam-sdks-java-io-mongodb', project ':beam-sdks-java-io-mqtt', project ':beam-sdks-java-io-redis', project ':beam-sdks-java-io-solr', project ':beam-sdks-java-io-tika', project ':beam-sdks-java-io-xml', project ':beam-sdks-java-javadoc', project ':beam-sdks-java-maven-archetypes-examples', project ':beam-sdks-java-maven-archetypes-starter', project ':beam-sdks-java-nexmark', project ':beam-sdks-python', project ':beam-sdks-python-container', project ':release']
Parallel execution with configuration on demand is an incubating feature.

> Configure project :
Evaluating root project 'beam' using build file '<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/build.gradle'.>
Applying build_rules.gradle to beam
Offline dependencies root configured at 'offline-repository'
createPerformanceTestHarness with default configuration for project beam
Adding 47 .gitignore exclusions to Apache Rat

> Configure project :beam-sdks-python
Evaluating project ':beam-sdks-python' using build file '<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/build.gradle'.>
Selected primary task ':beam-sdks-python:validatesRunnerTests' from project :beam-sdks-python
All projects evaluated.
Tasks to be executed: [task ':beam-sdks-python:setupVirtualenv', task ':beam-sdks-python:installGcpTest', task ':beam-sdks-python:validatesRunnerTests']
VisTEG: Dependency graph written into <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/build/reports/visteg.dot>
:beam-sdks-python:setupVirtualenv (Thread[Task worker for ':' Thread 6,5,main]) started.

> Task :beam-sdks-python:setupVirtualenv
Build cache key for task ':beam-sdks-python:setupVirtualenv' is 93b6db45447bce41b4b402e28c183369
Caching disabled for task ':beam-sdks-python:setupVirtualenv': Caching has not been enabled for the task
Task ':beam-sdks-python:setupVirtualenv' is not up-to-date because:
  No history is available.
Starting process 'command 'virtualenv''. Working directory: <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python> Command: virtualenv <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/build/gradleenv>
Successfully started process 'command 'virtualenv''
New python executable in <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/build/gradleenv/bin/python2>
Also creating executable in <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/build/gradleenv/bin/python>
Installing setuptools, pkg_resources, pip, wheel...done.
Running virtualenv with interpreter /usr/bin/python2
Starting process 'command 'sh''. Working directory: <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python> Command: sh -c . <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/build/gradleenv/bin/activate> && pip install --upgrade tox==3.0.0 grpcio-tools==1.3.5
Successfully started process 'command 'sh''
Collecting tox==3.0.0
  Using cached https://files.pythonhosted.org/packages/e6/41/4dcfd713282bf3213b0384320fa8841e4db032ddcb80bc08a540159d42a8/tox-3.0.0-py2.py3-none-any.whl
Collecting grpcio-tools==1.3.5
  Using cached https://files.pythonhosted.org/packages/05/f6/0296e29b1bac6f85d2a8556d48adf825307f73109a3c2c17fb734292db0a/grpcio_tools-1.3.5-cp27-cp27mu-manylinux1_x86_64.whl
Collecting pluggy<1.0,>=0.3.0 (from tox==3.0.0)
  Using cached https://files.pythonhosted.org/packages/82/05/43e3947125a2137cba4746135c75934ceed1863f27e050fc560052104a71/pluggy-0.6.0-py2-none-any.whl
Requirement not upgraded as not directly required: six in /usr/local/lib/python2.7/dist-packages (from tox==3.0.0) (1.11.0)
Requirement not upgraded as not directly required: virtualenv>=1.11.2 in /usr/lib/python2.7/dist-packages (from tox==3.0.0) (15.0.1)
Collecting py>=1.4.17 (from tox==3.0.0)
  Using cached https://files.pythonhosted.org/packages/67/a5/f77982214dd4c8fd104b066f249adea2c49e25e8703d284382eb5e9ab35a/py-1.5.3-py2.py3-none-any.whl
Collecting grpcio>=1.3.5 (from grpcio-tools==1.3.5)
  Using cached https://files.pythonhosted.org/packages/2a/d5/eb371e43d65989267a9daa90ee8f92d79b216184ea3c8cf0670e5c2388eb/grpcio-1.12.0-cp27-cp27mu-manylinux1_x86_64.whl
Collecting protobuf>=3.2.0 (from grpcio-tools==1.3.5)
  Using cached https://files.pythonhosted.org/packages/9d/61/54c3a9cfde6ffe0ca6a1786ddb8874263f4ca32e7693ad383bd8cf935015/protobuf-3.5.2.post1-cp27-cp27mu-manylinux1_x86_64.whl
Requirement not upgraded as not directly required: enum34>=1.0.4 in /usr/local/lib/python2.7/dist-packages (from grpcio>=1.3.5->grpcio-tools==1.3.5) (1.1.6)
Collecting futures>=2.2.0 (from grpcio>=1.3.5->grpcio-tools==1.3.5)
  Using cached https://files.pythonhosted.org/packages/2d/99/b2c4e9d5a30f6471e410a146232b4118e697fa3ffc06d6a65efde84debd0/futures-3.2.0-py2-none-any.whl
Requirement not upgraded as not directly required: setuptools in /usr/local/lib/python2.7/dist-packages (from protobuf>=3.2.0->grpcio-tools==1.3.5) (39.0.1)
Installing collected packages: pluggy, py, tox, futures, grpcio, protobuf, grpcio-tools
Could not install packages due to an EnvironmentError: [Errno 13] Permission denied: '/usr/local/lib/python2.7/dist-packages/pluggy-0.6.0.dist-info'
Consider using the `--user` option or check the permissions.


> Task :beam-sdks-python:setupVirtualenv FAILED
:beam-sdks-python:setupVirtualenv (Thread[Task worker for ':' Thread 6,5,main]) completed. Took 6.35 secs.

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/build.gradle'> line: 36

* What went wrong:
Execution failed for task ':beam-sdks-python:setupVirtualenv'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 11s
1 actionable task: 1 executed

Publishing build scan...
https://gradle.com/s/udhuyylju2yxu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

Jenkins build is back to normal : beam_PostCommit_Python_ValidatesRunner_Dataflow #1656

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1656/display/redirect?page=changes>


Build failed in Jenkins: beam_PostCommit_Python_ValidatesRunner_Dataflow #1655

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1655/display/redirect?page=changes>

Changes:

[Pablo] [BEAM-4249] Fix failing validates container test

[lcwik] [BEAM-4127] Flink streaming translation for portable pipeline. (#5358)

------------------------------------------
[...truncated 76.90 KB...]
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_singleton_with_different_defaults (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_singleton_without_unique_labels (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_default_value_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
  File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
  File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",> line 244, in test_par_do_with_multiple_outputs_and_using_return
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 102, in run
    result = super(TestPipeline, self).run(test_runner_api)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 389, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 402, in run
    return self.runner.run_pipeline(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 48, in run_pipeline
    self.result = super(TestDataflowRunner, self).run_pipeline(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 362, in run_pipeline
    self.dataflow_client.create_job(self.job), self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 180, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 470, in create_job
    self.create_job_description(job)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 500, in create_job_description
    job.options, file_copy=self._gcs_file_copy)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 329, in stage_job_resources
    setup_options.requirements_file, requirements_cache_path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 263, in _populate_requirements_cache
    processes.check_call(cmd_args)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",> line 44, in check_call
    return subprocess.check_call(*args, **kwargs)
  File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
    raise CalledProcessError(retcode, cmd)
CalledProcessError: Command '['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:']' returned non-zero exit status 2
-------------------- >> begin captured logging << --------------------
root: DEBUG: PValue computed by Some Numbers/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(some_fn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Create/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/ToVoidKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/pair_with_0 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/pair_with_1 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/GroupByKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Unkey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(some_fn) (tag odd): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Create/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/ToVoidKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/pair_with_1 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/pair_with_0 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/GroupByKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Unkey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(some_fn) (tag even): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Create/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/ToVoidKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/pair_with_0 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/pair_with_1 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/GroupByKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Unkey (tag None): refcount: 1 => 0
root: DEBUG: Connecting using Google Application Default Credentials.
root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-393926.1526429102.394223/pipeline.pb...
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-393926.1526429102.394223/pipeline.pb
root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-393926.1526429102.394223/requirements.txt...
root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-393926.1526429102.394223/requirements.txt
root: INFO: Executing command: ['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:']
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_flatten_multiple_pcollections_having_multiple_consumers (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",> line 558, in test_flatten_multiple_pcollections_having_multiple_consumers
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 102, in run
    result = super(TestPipeline, self).run(test_runner_api)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 389, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 402, in run
    return self.runner.run_pipeline(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 48, in run_pipeline
    self.result = super(TestDataflowRunner, self).run_pipeline(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 362, in run_pipeline
    self.dataflow_client.create_job(self.job), self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 180, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 470, in create_job
    self.create_job_description(job)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 500, in create_job_description
    job.options, file_copy=self._gcs_file_copy)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 329, in stage_job_resources
    setup_options.requirements_file, requirements_cache_path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 263, in _populate_requirements_cache
    processes.check_call(cmd_args)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",> line 44, in check_call
    return subprocess.check_call(*args, **kwargs)
  File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
    raise CalledProcessError(retcode, cmd)
CalledProcessError: Command '['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:']' returned non-zero exit status 2
-------------------- >> begin captured logging << --------------------
root: DEBUG: PValue computed by Start/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by Map(split_even_odd)/Map(split_even_odd) (tag even_length): refcount: 2 => 1
root: DEBUG: PValue computed by Map(split_even_odd)/Map(split_even_odd) (tag odd_length): refcount: 2 => 1
root: DEBUG: PValue computed by Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Create/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/ToVoidKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/pair_with_1 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/pair_with_0 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/GroupByKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert_that/Unkey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by Map(split_even_odd)/Map(split_even_odd) (tag even_length): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Create/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/ToVoidKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/pair_with_1 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/pair_with_0 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/GroupByKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:even/Unkey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by Map(split_even_odd)/Map(split_even_odd) (tag odd_length): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Create/Read (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/ToVoidKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/pair_with_0 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/pair_with_1 (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/Flatten (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/GroupByKey (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0
root: DEBUG: PValue computed by assert:odd/Unkey (tag None): refcount: 1 => 0
root: DEBUG: Connecting using Google Application Default Credentials.
root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-011370.1526429102.011696/pipeline.pb...
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-011370.1526429102.011696/pipeline.pb
root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-011370.1526429102.011696/requirements.txt...
root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0516000502-011370.1526429102.011696/requirements.txt
root: INFO: Executing command: ['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:']
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 16 tests in 1815.785s

FAILED (errors=2)
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_05_15-14956991075761740292?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_11_53-11464754543387914824?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_19_30-16834578744107121031?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_26_56-18007238553591732999?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_05_19-2537501766015507418?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_13_32-7582102487891854126?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_20_39-14969698208971532080?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_05_14-13093430216925216080?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_13_31-12006268181024095829?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_20_08-4878531754279132335?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_27_29-11119682368754213266?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_05_13-15802834407269003019?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_14_05-11476759468816629312?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_17_20_58-10094282950362579833?project=apache-beam-testing.
Build step 'Execute shell' marked build as failure