You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2017/05/09 21:05:51 UTC

Build failed in Jenkins: beam_PostCommit_Python_Verify #2170

See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2170/display/redirect>

------------------------------------------
[...truncated 545.07 KB...]
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 396, in save_reduce
    save(cls)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 92, in wrapper
    obj=obj)
  File "/usr/lib/python2.7/pickle.py", line 401, in save_reduce
    save(args)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "/usr/lib/python2.7/pickle.py", line 562, in save_tuple
    save(element)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 644, in save_dict
    write(EMPTY_DICT)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 177, in test_iterable_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 162, in run
    if test_runner_api and self._verify_runner_api_compatible():
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 332, in _verify_runner_api_compatible
    self.visit(Visitor())
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 196, in visit
    self._root_transform().visit(visitor, self, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 489, in visit
    part.visit(visitor, pipeline, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 489, in visit
    part.visit(visitor, pipeline, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 489, in visit
    part.visit(visitor, pipeline, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 492, in visit
    visitor.visit_transform(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 324, in visit_transform
    enable_trace=False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 215, in loads
    return dill.loads(s)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 277, in loads
    return load(file)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 266, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1190, in load_appends
    del stack[mark:]
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 156, in test_multi_valued_singleton_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 35, in run
    self.result = super(TestDataflowRunner, self).run(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 249, in run
    self.dataflow_client.create_job(self.job), self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 423, in create_job
    self.create_job_description(job)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 446, in create_job_description
    job.options, file_copy=self._gcs_file_copy)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 294, in stage_job_resources
    file_copy(setup_options.requirements_file, staged_path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 386, in _gcs_file_copy
    self.stage_file(to_folder, to_name, f)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 405, in stage_file
    response = self._storage_client.objects.Insert(request, upload=upload)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py",> line 971, in Insert
    download=download)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 721, in _RunMethod
    http, http_request, **opts)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 341, in MakeRequest
    check_response_func=check_response_func)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 391, in _MakeRequestNoRetry
    redirections=redirections, connection_type=connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/oauth2client/transport.py",> line 169, in new_request
    redirections, connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1609, in request
    (response, content) = self._request(conn, authority, uri, request_uri, method, body, headers, redirections, cachekey)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1351, in _request
    (response, content) = self._conn_request(conn, request_uri, method, body, headers)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1307, in _conn_request
    response = conn.getresponse()
  File "/usr/lib/python2.7/httplib.py", line 1089, in getresponse
    response.begin()
  File "/usr/lib/python2.7/httplib.py", line 444, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python2.7/httplib.py", line 400, in _read_status
    line = self.fp.readline(_MAXLINE + 1)
  File "/usr/lib/python2.7/socket.py", line 476, in readline
    data = self._sock.recv(self._rbufsize)
  File "/usr/lib/python2.7/ssl.py", line 341, in recv
    return self.read(buflen)
  File "/usr/lib/python2.7/ssl.py", line 260, in read
    return self._sslobj.read(len)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 235, in test_par_do_with_multiple_outputs_and_using_return
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)'

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 219, in test_par_do_with_multiple_outputs_and_using_yield
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest)'

======================================================================
ERROR: test_read_from_text_metrics (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 190, in test_read_from_text_metrics
    res = pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_read_from_text_metrics (apache_beam.transforms.ptransform_test.PTransformTest)'

----------------------------------------------------------------------
Ran 15 tests in 902.610s

FAILED (errors=15)
Build step 'Execute shell' marked build as failure

Jenkins build is back to normal : beam_PostCommit_Python_Verify #2179

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2179/display/redirect?page=changes>


Build failed in Jenkins: beam_PostCommit_Python_Verify #2178

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2178/display/redirect?page=changes>

Changes:

[chamikara] Mark PipelineVisitor and AppliedPTransform as internal.

------------------------------------------
[...truncated 583.17 KB...]
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_merge_tagged_vals_under_key"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key).out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s11"
        }, 
        "serialized_fn": "<string of 1344 bytes>", 
        "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s13", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Unkey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s12"
        }, 
        "serialized_fn": "<string of 964 bytes>", 
        "user_name": "assert_that/Unkey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s14", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s13"
        }, 
        "serialized_fn": "<string of 1116 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 10 May 2017 07:22:30 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(fb22bb4935c0dc05): The workflow could not be created. Causes: (6df4931edcda8554): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 10 May 2017 07:22:32 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(977b16180b57a7b6): The workflow could not be created. Causes: (e74ec17bc93ce3ec): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '440', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 10 May 2017 07:22:37 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(b9324ad1166a5c08): The workflow could not be created. Causes: (2f56322beac9b6e): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 10 May 2017 07:22:48 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(6558385183e10a00): The workflow could not be created. Causes: (402d72dae45e3644): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: ERROR: HTTP status 429 trying to create job at dataflow service endpoint https://dataflow.googleapis.com
root: CRITICAL: details of server error: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 10 May 2017 07:23:07 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(f61fa0d40c324813): The workflow could not be created. Causes: (85181ea36db4428f): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 15 tests in 1482.331s

FAILED (errors=1)
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_09_51-14322785757164930721?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_16_08-14527735254560773830?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_22_20-16513378811748848808?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_27_49-11934288747770486948?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_09_48-12583129661156031828?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_16_25-5994599331681716108?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_22_15-9143139617157447675?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_28_21-7513777183059118878?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_09_48-12909723208104283468?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_17_03-8512959352195195629?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_23_13-10107273846410910154?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_09_48-8292473655447987501?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_16_44-5863574620033422371?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-10_00_23_25-3291124431490556190?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Not sending mail to unregistered user kirpichov@google.com
Not sending mail to unregistered user ccy@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2177

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2177/display/redirect?page=changes>

Changes:

[tgroh] Mark PValue and PValueBase Internal

------------------------------------------
[...truncated 1.76 MB...]
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s14", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s13"
        }, 
        "serialized_fn": "<string of 1120 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2017-05-10T04:43:52.075472Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2017-05-09_21_43_51-2632386391657042642'
 location: u'global'
 name: u'beamapp-jenkins-0510044337-649531'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-05-09_21_43_51-2632386391657042642]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-05-09_21_43_51-2632386391657042642
root: INFO: Job 2017-05-09_21_43_51-2632386391657042642 is in state JOB_STATE_RUNNING
root: INFO: 2017-05-10T04:43:51.632Z: JOB_MESSAGE_WARNING: (24881d79ea4a53be): Setting the number of workers (1) disables autoscaling for this job. If you are trying to cap autoscaling, consider only setting max_num_workers. If you want to disable autoscaling altogether, the documented way is to explicitly use autoscalingAlgorithm=NONE.
root: INFO: 2017-05-10T04:43:52.695Z: JOB_MESSAGE_DETAILED: (cf18b6e5c174977b): Checking required Cloud APIs are enabled.
root: INFO: 2017-05-10T04:43:54.885Z: JOB_MESSAGE_DEBUG: (cf18b6e5c174915c): Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2017-05-10T04:43:54.888Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749b66): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-05-10T04:43:54.890Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749570): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-05-10T04:43:54.895Z: JOB_MESSAGE_DEBUG: (cf18b6e5c1749984): Annotating graph with Autotuner information.
root: INFO: 2017-05-10T04:43:54.908Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749d98): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-05-10T04:43:54.917Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749bb6): Unzipping flatten s10 for input s8.out
root: INFO: 2017-05-10T04:43:54.920Z: JOB_MESSAGE_DETAILED: (cf18b6e5c17495c0): Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten , into producer assert_that/Group/pair_with_0
root: INFO: 2017-05-10T04:43:54.923Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749fca): Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2017-05-10T04:43:54.925Z: JOB_MESSAGE_DETAILED: (cf18b6e5c17499d4): Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2017-05-10T04:43:54.930Z: JOB_MESSAGE_DETAILED: (cf18b6e5c17493de): Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2017-05-10T04:43:54.937Z: JOB_MESSAGE_DETAILED: (cf18b6e5c17497f2): Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read
root: INFO: 2017-05-10T04:43:54.945Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749610): Unzipping flatten s10-u13 for input s11-reify-value0-c11
root: INFO: 2017-05-10T04:43:54.949Z: JOB_MESSAGE_DETAILED: (cf18b6e5c174901a): Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten , into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T04:43:54.951Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749a24): Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2017-05-10T04:43:54.953Z: JOB_MESSAGE_DETAILED: (cf18b6e5c174942e): Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T04:43:54.957Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749e38): Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2017-05-10T04:43:54.959Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749842): Fusing consumer assert_that/WindowInto(WindowIntoFn) into compute/Do
root: INFO: 2017-05-10T04:43:54.961Z: JOB_MESSAGE_DETAILED: (cf18b6e5c174924c): Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2017-05-10T04:43:54.964Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749c56): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2017-05-10T04:43:54.966Z: JOB_MESSAGE_DETAILED: (cf18b6e5c1749660): Fusing consumer compute/Do into start/Read
root: INFO: 2017-05-10T04:43:55.026Z: JOB_MESSAGE_DEBUG: (cf18b6e5c1749750): Workflow config is missing a default resource spec.
root: INFO: 2017-05-10T04:43:55.028Z: JOB_MESSAGE_DETAILED: (cf18b6e5c174915a): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-05-10T04:43:55.031Z: JOB_MESSAGE_DEBUG: (cf18b6e5c1749b64): Adding workflow start and stop steps.
root: INFO: 2017-05-10T04:43:55.033Z: JOB_MESSAGE_DEBUG: (cf18b6e5c174956e): Assigning stage ids.
root: INFO: 2017-05-10T04:43:55.073Z: JOB_MESSAGE_DEBUG: (d010e88f11ba2266): Executing wait step start21
root: INFO: 2017-05-10T04:43:55.081Z: JOB_MESSAGE_BASIC: (d010e88f11ba2e8d): Executing operation side/Read
root: INFO: 2017-05-10T04:43:55.092Z: JOB_MESSAGE_DEBUG: (d010e88f11ba29ef): Value "side/Read.out" materialized.
root: INFO: 2017-05-10T04:43:55.101Z: JOB_MESSAGE_BASIC: (d010e88f11ba2616): Executing operation compute/SideInput-s3
root: INFO: 2017-05-10T04:43:55.111Z: JOB_MESSAGE_DEBUG: (d010e88f11ba2178): Value "compute/SideInput-s3.output" materialized.
root: INFO: 2017-05-10T04:43:55.117Z: JOB_MESSAGE_BASIC: (d010e88f11ba2cda): Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2017-05-10T04:43:55.323Z: JOB_MESSAGE_DEBUG: (51e28024bc9536aa): Starting worker pool setup.
root: INFO: 2017-05-10T04:43:55.325Z: JOB_MESSAGE_BASIC: (51e28024bc95309c): Starting 1 workers...
root: INFO: 2017-05-10T04:43:55.340Z: JOB_MESSAGE_DEBUG: (d010e88f11ba25ed): Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2017-05-10T04:43:55.351Z: JOB_MESSAGE_BASIC: (d010e88f11ba2fc5): Executing operation start/Read+compute/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T04:43:55.361Z: JOB_MESSAGE_BASIC: (9aa9701bfc6ae545): Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T04:45:02.352Z: JOB_MESSAGE_DETAILED: (32554706c3c5c9c1): Workers have started successfully.
root: INFO: 2017-05-10T04:47:46.860Z: JOB_MESSAGE_ERROR: (8343e7f25c211a5d): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:47:50.231Z: JOB_MESSAGE_ERROR: (8343e7f25c211ed0): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:47:53.622Z: JOB_MESSAGE_ERROR: (8343e7f25c211343): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:47:56.962Z: JOB_MESSAGE_ERROR: (8343e7f25c211913): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:48:00.321Z: JOB_MESSAGE_ERROR: (8343e7f25c211ee3): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:48:03.693Z: JOB_MESSAGE_ERROR: (8343e7f25c2114b3): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:48:07.082Z: JOB_MESSAGE_ERROR: (8343e7f25c211926): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:48:10.512Z: JOB_MESSAGE_ERROR: (8343e7f25c211d99): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:48:13.934Z: JOB_MESSAGE_ERROR: (8343e7f25c21120c): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T04:48:14.081Z: JOB_MESSAGE_DEBUG: (d010e88f11ba274e): Executing failure step failure20
root: INFO: 2017-05-10T04:48:14.083Z: JOB_MESSAGE_ERROR: (d010e88f11ba25c4): Workflow failed. Causes: (d010e88f11ba2a62): S05:start/Read+compute/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write failed.
root: INFO: 2017-05-10T04:48:14.140Z: JOB_MESSAGE_DETAILED: (cf18b6e5c17494a5): Cleaning up.
root: INFO: 2017-05-10T04:48:14.270Z: JOB_MESSAGE_DEBUG: (cf18b6e5c1749eaf): Starting worker pool teardown.
root: INFO: 2017-05-10T04:48:14.272Z: JOB_MESSAGE_BASIC: (cf18b6e5c17498b9): Stopping worker pool...
root: INFO: 2017-05-10T04:49:39.278Z: JOB_MESSAGE_BASIC: (cf18b6e5c1749fc8): Worker pool stopped.
root: INFO: 2017-05-10T04:49:39.325Z: JOB_MESSAGE_DEBUG: (cf18b6e5c17494f5): Tearing down pending resources...
root: INFO: Job 2017-05-09_21_43_51-2632386391657042642 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 15 tests in 1444.998s

FAILED (errors=14)
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_26_43-7494046603359720390?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_33_05-17778654726138252063?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_39_35-1019057384112298327?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_26_43-2743303429761940666?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_32_15-9014687942684870733?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_38_26-5769702013089870763?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_43_51-2632386391657042642?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_26_43-8830841609316196593?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_32_06-878169968002913237?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_38_22-13458795397275563733?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_45_16-14516527290768753107?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_26_41-18175790497212241534?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_32_43-10126553643447272512?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_38_08-2485544320882357778?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_21_43_29-12045478608446743663?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Not sending mail to unregistered user kirpichov@google.com
Not sending mail to unregistered user ccy@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2176

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2176/display/redirect>

------------------------------------------
[...truncated 1.75 MB...]
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Unkey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s12"
        }, 
        "serialized_fn": "<string of 964 bytes>", 
        "user_name": "assert_that/Unkey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s14", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s13"
        }, 
        "serialized_fn": "<string of 1120 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2017-05-10T03:33:20.693317Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2017-05-09_20_33_20-14479349599173966395'
 location: u'global'
 name: u'beamapp-jenkins-0510033306-296809'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-05-09_20_33_20-14479349599173966395]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-05-09_20_33_20-14479349599173966395
root: INFO: Job 2017-05-09_20_33_20-14479349599173966395 is in state JOB_STATE_RUNNING
root: INFO: 2017-05-10T03:33:20.159Z: JOB_MESSAGE_WARNING: (c8f0fbbedd757e2f): Setting the number of workers (1) disables autoscaling for this job. If you are trying to cap autoscaling, consider only setting max_num_workers. If you want to disable autoscaling altogether, the documented way is to explicitly use autoscalingAlgorithm=NONE.
root: INFO: 2017-05-10T03:33:21.409Z: JOB_MESSAGE_DETAILED: (d1eb603b16feebaa): Checking required Cloud APIs are enabled.
root: INFO: 2017-05-10T03:33:23.486Z: JOB_MESSAGE_DEBUG: (d1eb603b16feeed4): Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2017-05-10T03:33:23.488Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee53a): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-05-10T03:33:23.491Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeba0): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-05-10T03:33:23.496Z: JOB_MESSAGE_DEBUG: (d1eb603b16fee86c): Annotating graph with Autotuner information.
root: INFO: 2017-05-10T03:33:23.508Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee538): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-05-10T03:33:23.516Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee86a): Unzipping flatten s10 for input s8.out
root: INFO: 2017-05-10T03:33:23.518Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeed0): Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten , into producer assert_that/Group/pair_with_0
root: INFO: 2017-05-10T03:33:23.520Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee536): Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2017-05-10T03:33:23.523Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeb9c): Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2017-05-10T03:33:23.525Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee202): Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2017-05-10T03:33:23.531Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeece): Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read
root: INFO: 2017-05-10T03:33:23.538Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee200): Unzipping flatten s10-u13 for input s11-reify-value0-c11
root: INFO: 2017-05-10T03:33:23.540Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee866): Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten , into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T03:33:23.542Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeecc): Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2017-05-10T03:33:23.544Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee532): Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T03:33:23.547Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeb98): Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2017-05-10T03:33:23.549Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee1fe): Fusing consumer assert_that/WindowInto(WindowIntoFn) into compute/Do
root: INFO: 2017-05-10T03:33:23.551Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee864): Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2017-05-10T03:33:23.554Z: JOB_MESSAGE_DETAILED: (d1eb603b16feeeca): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2017-05-10T03:33:23.556Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee530): Fusing consumer compute/Do into start/Read
root: INFO: 2017-05-10T03:33:23.615Z: JOB_MESSAGE_DEBUG: (d1eb603b16feeec0): Workflow config is missing a default resource spec.
root: INFO: 2017-05-10T03:33:23.617Z: JOB_MESSAGE_DETAILED: (d1eb603b16fee526): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-05-10T03:33:23.619Z: JOB_MESSAGE_DEBUG: (d1eb603b16feeb8c): Adding workflow start and stop steps.
root: INFO: 2017-05-10T03:33:23.622Z: JOB_MESSAGE_DEBUG: (d1eb603b16fee1f2): Assigning stage ids.
root: INFO: 2017-05-10T03:33:23.653Z: JOB_MESSAGE_DEBUG: (cdb940e1e12b627e): Executing wait step start21
root: INFO: 2017-05-10T03:33:23.660Z: JOB_MESSAGE_BASIC: (cdb940e1e12b6c5b): Executing operation side/Read
root: INFO: 2017-05-10T03:33:23.671Z: JOB_MESSAGE_DEBUG: (cdb940e1e12b6831): Value "side/Read.out" materialized.
root: INFO: 2017-05-10T03:33:23.677Z: JOB_MESSAGE_BASIC: (cdb940e1e12b620e): Executing operation compute/SideInput-s3
root: INFO: 2017-05-10T03:33:23.686Z: JOB_MESSAGE_DEBUG: (cdb940e1e12b6de4): Value "compute/SideInput-s3.output" materialized.
root: INFO: 2017-05-10T03:33:23.694Z: JOB_MESSAGE_BASIC: (cdb940e1e12b69ba): Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2017-05-10T03:33:23.897Z: JOB_MESSAGE_DEBUG: (8897929db08c710e): Starting worker pool setup.
root: INFO: 2017-05-10T03:33:23.899Z: JOB_MESSAGE_BASIC: (8897929db08c7d30): Starting 1 workers...
root: INFO: 2017-05-10T03:33:23.918Z: JOB_MESSAGE_DEBUG: (cdb940e1e12b6b7b): Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2017-05-10T03:33:23.931Z: JOB_MESSAGE_BASIC: (cdb940e1e12b6b43): Executing operation start/Read+compute/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T03:33:23.934Z: JOB_MESSAGE_BASIC: (4f249eccba3e87f7): Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T03:35:34.795Z: JOB_MESSAGE_DETAILED: (f26b9262b108cd47): Workers have started successfully.
root: INFO: 2017-05-10T03:37:30.948Z: JOB_MESSAGE_ERROR: (f8cdea404c6a0cfa): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:34.321Z: JOB_MESSAGE_ERROR: (f8cdea404c6a0f39): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:37.686Z: JOB_MESSAGE_ERROR: (f8cdea404c6a0178): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:41.010Z: JOB_MESSAGE_ERROR: (f8cdea404c6a0e88): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:44.343Z: JOB_MESSAGE_ERROR: (f8cdea404c6a0b98): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:46.676Z: JOB_MESSAGE_ERROR: (f8cdea404c6a08a8): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:50.041Z: JOB_MESSAGE_ERROR: (f8cdea404c6a05b8): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T03:37:50.063Z: JOB_MESSAGE_DEBUG: (4f249eccba3e8b34): Executing failure step failure20
root: INFO: 2017-05-10T03:37:50.066Z: JOB_MESSAGE_ERROR: (4f249eccba3e88e2): Workflow failed. Causes: (4f249eccba3e8fd8): S04:assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write failed.
root: INFO: 2017-05-10T03:37:50.123Z: JOB_MESSAGE_DETAILED: (d1eb603b16feed23): Cleaning up.
root: INFO: 2017-05-10T03:37:50.255Z: JOB_MESSAGE_DEBUG: (d1eb603b16fee389): Starting worker pool teardown.
root: INFO: 2017-05-10T03:37:50.257Z: JOB_MESSAGE_BASIC: (d1eb603b16fee9ef): Stopping worker pool...
root: INFO: 2017-05-10T03:39:00.270Z: JOB_MESSAGE_BASIC: (d1eb603b16feeb88): Worker pool stopped.
root: INFO: 2017-05-10T03:39:00.314Z: JOB_MESSAGE_DEBUG: (d1eb603b16feed1f): Tearing down pending resources...
root: INFO: Job 2017-05-09_20_33_20-14479349599173966395 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 15 tests in 1451.855s

FAILED (errors=14)
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_15_13-12643813889606878861?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_20_54-11062214958168050858?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_27_18-462380442472603986?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_33_22-12004211461037347708?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_15_11-15410755755379746035?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_21_31-7572223232558927677?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_27_25-13403432425446460475?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_33_20-14479349599173966395?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_15_11-14829333216387362712?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_21_36-15029687597997686332?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_27_22-6913675982676463131?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_15_11-4225663317918682187?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_20_55-11381664972944263137?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_27_10-12809840961614428869?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_20_32_25-6627863476686600574?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Not sending mail to unregistered user kirpichov@google.com
Not sending mail to unregistered user ccy@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2175

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2175/display/redirect?page=changes>

Changes:

[altay] [BEAM-2236] Move test utilities out of python core

------------------------------------------
[...truncated 1.74 MB...]
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Unkey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s14"
        }, 
        "serialized_fn": "<string of 964 bytes>", 
        "user_name": "assert_that/Unkey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s16", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s15"
        }, 
        "serialized_fn": "<string of 1120 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2017-05-10T02:19:23.071895Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2017-05-09_19_19_22-8297893128687954825'
 location: u'global'
 name: u'beamapp-jenkins-0510021902-159349'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-05-09_19_19_22-8297893128687954825]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-05-09_19_19_22-8297893128687954825
root: INFO: Job 2017-05-09_19_19_22-8297893128687954825 is in state JOB_STATE_RUNNING
root: INFO: 2017-05-10T02:19:22.469Z: JOB_MESSAGE_WARNING: (732809d61f99ae9d): Setting the number of workers (1) disables autoscaling for this job. If you are trying to cap autoscaling, consider only setting max_num_workers. If you want to disable autoscaling altogether, the documented way is to explicitly use autoscalingAlgorithm=NONE.
root: INFO: 2017-05-10T02:19:23.618Z: JOB_MESSAGE_DETAILED: (215f06164378b15b): Checking required Cloud APIs are enabled.
root: INFO: 2017-05-10T02:19:25.777Z: JOB_MESSAGE_DEBUG: (215f06164378b184): Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2017-05-10T02:19:25.779Z: JOB_MESSAGE_DETAILED: (215f06164378b45e): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-05-10T02:19:25.782Z: JOB_MESSAGE_DETAILED: (215f06164378b738): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-05-10T02:19:25.786Z: JOB_MESSAGE_DEBUG: (215f06164378bcec): Annotating graph with Autotuner information.
root: INFO: 2017-05-10T02:19:25.798Z: JOB_MESSAGE_DETAILED: (215f06164378b2a0): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-05-10T02:19:25.806Z: JOB_MESSAGE_DETAILED: (215f06164378bb2e): Unzipping flatten s12 for input s10.out
root: INFO: 2017-05-10T02:19:25.809Z: JOB_MESSAGE_DETAILED: (215f06164378be08): Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten , into producer assert_that/Group/pair_with_0
root: INFO: 2017-05-10T02:19:25.814Z: JOB_MESSAGE_DETAILED: (215f06164378b3bc): Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2017-05-10T02:19:25.817Z: JOB_MESSAGE_DETAILED: (215f06164378b696): Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2017-05-10T02:19:25.819Z: JOB_MESSAGE_DETAILED: (215f06164378b970): Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read
root: INFO: 2017-05-10T02:19:25.822Z: JOB_MESSAGE_DETAILED: (215f06164378bc4a): Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2017-05-10T02:19:25.833Z: JOB_MESSAGE_DETAILED: (215f06164378b4d8): Unzipping flatten s12-u13 for input s13-reify-value0-c11
root: INFO: 2017-05-10T02:19:25.836Z: JOB_MESSAGE_DETAILED: (215f06164378b7b2): Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten , into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T02:19:25.841Z: JOB_MESSAGE_DETAILED: (215f06164378ba8c): Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2017-05-10T02:19:25.855Z: JOB_MESSAGE_DETAILED: (215f06164378b8ce): Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T02:19:25.859Z: JOB_MESSAGE_DETAILED: (215f06164378bba8): Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2017-05-10T02:19:25.862Z: JOB_MESSAGE_DETAILED: (215f06164378be82): Fusing consumer FlatMap(<lambda at sideinputs_test.py:301>)/Do into main input/Read
root: INFO: 2017-05-10T02:19:25.865Z: JOB_MESSAGE_DETAILED: (215f06164378b15c): Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2017-05-10T02:19:25.875Z: JOB_MESSAGE_DETAILED: (215f06164378b436): Fusing consumer assert_that/WindowInto(WindowIntoFn) into FlatMap(<lambda at sideinputs_test.py:301>)/Do
root: INFO: 2017-05-10T02:19:25.880Z: JOB_MESSAGE_DETAILED: (215f06164378b710): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2017-05-10T02:19:25.954Z: JOB_MESSAGE_DEBUG: (215f06164378b40e): Workflow config is missing a default resource spec.
root: INFO: 2017-05-10T02:19:25.956Z: JOB_MESSAGE_DETAILED: (215f06164378b6e8): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-05-10T02:19:25.958Z: JOB_MESSAGE_DEBUG: (215f06164378b9c2): Adding workflow start and stop steps.
root: INFO: 2017-05-10T02:19:25.960Z: JOB_MESSAGE_DEBUG: (215f06164378bc9c): Assigning stage ids.
root: INFO: 2017-05-10T02:19:25.994Z: JOB_MESSAGE_DEBUG: (b1fd05e9a7bedd47): Executing wait step start22
root: INFO: 2017-05-10T02:19:26.002Z: JOB_MESSAGE_BASIC: (b1fd05e9a7bed5e1): Executing operation side1/Read
root: INFO: 2017-05-10T02:19:26.004Z: JOB_MESSAGE_BASIC: (c118619b16e5d0d0): Executing operation side2/Read
root: INFO: 2017-05-10T02:19:26.011Z: JOB_MESSAGE_DEBUG: (b1fd05e9a7bede7b): Value "side1/Read.out" materialized.
root: INFO: 2017-05-10T02:19:26.017Z: JOB_MESSAGE_DEBUG: (c118619b16e5d6f2): Value "side2/Read.out" materialized.
root: INFO: 2017-05-10T02:19:26.025Z: JOB_MESSAGE_BASIC: (c118619b16e5deb9): Executing operation Flatten
root: INFO: 2017-05-10T02:19:26.037Z: JOB_MESSAGE_DEBUG: (c118619b16e5de47): Value "Flatten.out" materialized.
root: INFO: 2017-05-10T02:19:26.045Z: JOB_MESSAGE_BASIC: (c118619b16e5d8e6): Executing operation FlatMap(<lambda at sideinputs_test.py:301>)/SideInput-s5
root: INFO: 2017-05-10T02:19:26.057Z: JOB_MESSAGE_DEBUG: (c118619b16e5d1e0): Value "FlatMap(<lambda at sideinputs_test.py:301>)/SideInput-s5.output" materialized.
root: INFO: 2017-05-10T02:19:26.066Z: JOB_MESSAGE_BASIC: (c118619b16e5dada): Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2017-05-10T02:19:26.270Z: JOB_MESSAGE_DEBUG: (e2a07eb64d8d91dc): Starting worker pool setup.
root: INFO: 2017-05-10T02:19:26.272Z: JOB_MESSAGE_BASIC: (e2a07eb64d8d9a8e): Starting 1 workers...
root: INFO: 2017-05-10T02:19:26.289Z: JOB_MESSAGE_DEBUG: (c118619b16e5dc0d): Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2017-05-10T02:19:26.300Z: JOB_MESSAGE_BASIC: (c118619b16e5dee5): Executing operation main input/Read+FlatMap(<lambda at sideinputs_test.py:301>)/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T02:19:26.304Z: JOB_MESSAGE_BASIC: (b1fd05e9a7bed11e): Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T02:21:48.042Z: JOB_MESSAGE_DETAILED: (3da1880b746fd8d): Workers have started successfully.
root: INFO: 2017-05-10T02:23:52.144Z: JOB_MESSAGE_ERROR: (19edd31651e8eddb): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T02:23:55.557Z: JOB_MESSAGE_ERROR: (19edd31651e8ec23): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T02:23:58.987Z: JOB_MESSAGE_ERROR: (19edd31651e8ea6b): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T02:24:02.382Z: JOB_MESSAGE_ERROR: (19edd31651e8e8b3): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T02:24:02.405Z: JOB_MESSAGE_DEBUG: (c118619b16e5d63a): Executing failure step failure21
root: INFO: 2017-05-10T02:24:02.407Z: JOB_MESSAGE_ERROR: (c118619b16e5d2f0): Workflow failed. Causes: (c118619b16e5d362): S07:main input/Read+FlatMap(<lambda at sideinputs_test.py:301>)/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write failed.
root: INFO: 2017-05-10T02:24:02.465Z: JOB_MESSAGE_DETAILED: (215f06164378bc4b): Cleaning up.
root: INFO: 2017-05-10T02:24:02.597Z: JOB_MESSAGE_DEBUG: (215f06164378bf25): Starting worker pool teardown.
root: INFO: 2017-05-10T02:24:02.599Z: JOB_MESSAGE_BASIC: (215f06164378b1ff): Stopping worker pool...
root: INFO: 2017-05-10T02:25:27.612Z: JOB_MESSAGE_BASIC: (215f06164378b646): Worker pool stopped.
root: INFO: 2017-05-10T02:25:27.644Z: JOB_MESSAGE_DEBUG: (215f06164378b31b): Tearing down pending resources...
root: INFO: Job 2017-05-09_19_19_22-8297893128687954825 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 15 tests in 1479.364s

FAILED (errors=14)
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_01_21-248271457926087266?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_07_47-8258461548859220162?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_14_08-14453897031660960648?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_20_02-12788238450635658753?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_01_20-13230321987298668929?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_08_12-17925044393466132278?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_14_09-9609217760774917720?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_01_23-18236196026040646881?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_07_48-8000552664810130483?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_13_25-2575260572870413036?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_19_22-8297893128687954825?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_01_19-11605424239888703192?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_08_27-2672674185206199168?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_14_19-3970926101876107287?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_19_19_33-5387510595643630118?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Not sending mail to unregistered user kirpichov@google.com
Not sending mail to unregistered user ccy@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2174

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2174/display/redirect?page=changes>

Changes:

[ccy] Update Apache Beam Python version to 2.1.0.dev

[davor] Register TestSparkPipelineOptions only in src/test to avoid hard

[davor] Add per-runner profile to Java 8 examples

[altay] Remove trailing whitespace

[davor] Shade dependencies in sdks/core

[davor] Fix checkstyle error

[davor] Include 'sun.reflect' in GcpCoreApiSurfaceTest

------------------------------------------
[...truncated 1.75 MB...]
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Unkey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s12"
        }, 
        "serialized_fn": "<string of 964 bytes>", 
        "user_name": "assert_that/Unkey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s14", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s13"
        }, 
        "serialized_fn": "<string of 1120 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2017-05-10T01:38:20.116590Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2017-05-09_18_38_19-10569987656720048055'
 location: u'global'
 name: u'beamapp-jenkins-0510013805-731745'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-05-09_18_38_19-10569987656720048055]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-05-09_18_38_19-10569987656720048055
root: INFO: Job 2017-05-09_18_38_19-10569987656720048055 is in state JOB_STATE_RUNNING
root: INFO: 2017-05-10T01:38:19.388Z: JOB_MESSAGE_WARNING: (92b023c5870fd5b3): Setting the number of workers (1) disables autoscaling for this job. If you are trying to cap autoscaling, consider only setting max_num_workers. If you want to disable autoscaling altogether, the documented way is to explicitly use autoscalingAlgorithm=NONE.
root: INFO: 2017-05-10T01:38:21.145Z: JOB_MESSAGE_DETAILED: (28e4844b147a2c77): Checking required Cloud APIs are enabled.
root: INFO: 2017-05-10T01:38:23.634Z: JOB_MESSAGE_DEBUG: (28e4844b147a2265): Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2017-05-10T01:38:23.636Z: JOB_MESSAGE_DETAILED: (28e4844b147a22a7): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-05-10T01:38:23.639Z: JOB_MESSAGE_DETAILED: (28e4844b147a22e9): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-05-10T01:38:23.643Z: JOB_MESSAGE_DEBUG: (28e4844b147a236d): Annotating graph with Autotuner information.
root: INFO: 2017-05-10T01:38:23.655Z: JOB_MESSAGE_DETAILED: (28e4844b147a23f1): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-05-10T01:38:23.663Z: JOB_MESSAGE_DETAILED: (28e4844b147a24b7): Unzipping flatten s10 for input s8.out
root: INFO: 2017-05-10T01:38:23.666Z: JOB_MESSAGE_DETAILED: (28e4844b147a24f9): Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten , into producer assert_that/Group/pair_with_0
root: INFO: 2017-05-10T01:38:23.669Z: JOB_MESSAGE_DETAILED: (28e4844b147a253b): Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2017-05-10T01:38:23.671Z: JOB_MESSAGE_DETAILED: (28e4844b147a257d): Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2017-05-10T01:38:23.673Z: JOB_MESSAGE_DETAILED: (28e4844b147a25bf): Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2017-05-10T01:38:23.679Z: JOB_MESSAGE_DETAILED: (28e4844b147a2643): Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read
root: INFO: 2017-05-10T01:38:23.686Z: JOB_MESSAGE_DETAILED: (28e4844b147a2709): Unzipping flatten s10-u13 for input s11-reify-value0-c11
root: INFO: 2017-05-10T01:38:23.689Z: JOB_MESSAGE_DETAILED: (28e4844b147a274b): Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten , into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T01:38:23.691Z: JOB_MESSAGE_DETAILED: (28e4844b147a278d): Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2017-05-10T01:38:23.693Z: JOB_MESSAGE_DETAILED: (28e4844b147a27cf): Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2017-05-10T01:38:23.695Z: JOB_MESSAGE_DETAILED: (28e4844b147a2811): Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2017-05-10T01:38:23.698Z: JOB_MESSAGE_DETAILED: (28e4844b147a2853): Fusing consumer assert_that/WindowInto(WindowIntoFn) into compute/Do
root: INFO: 2017-05-10T01:38:23.701Z: JOB_MESSAGE_DETAILED: (28e4844b147a2895): Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2017-05-10T01:38:23.703Z: JOB_MESSAGE_DETAILED: (28e4844b147a28d7): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2017-05-10T01:38:23.706Z: JOB_MESSAGE_DETAILED: (28e4844b147a2919): Fusing consumer compute/Do into start/Read
root: INFO: 2017-05-10T01:38:23.759Z: JOB_MESSAGE_DEBUG: (28e4844b147a2f49): Workflow config is missing a default resource spec.
root: INFO: 2017-05-10T01:38:23.761Z: JOB_MESSAGE_DETAILED: (28e4844b147a2f8b): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-05-10T01:38:23.763Z: JOB_MESSAGE_DEBUG: (28e4844b147a2fcd): Adding workflow start and stop steps.
root: INFO: 2017-05-10T01:38:23.765Z: JOB_MESSAGE_DEBUG: (28e4844b147a200f): Assigning stage ids.
root: INFO: 2017-05-10T01:38:23.795Z: JOB_MESSAGE_DEBUG: (6d11656c1e5f4c23): Executing wait step start21
root: INFO: 2017-05-10T01:38:23.802Z: JOB_MESSAGE_BASIC: (6d11656c1e5f4694): Executing operation side/Read
root: INFO: 2017-05-10T01:38:23.811Z: JOB_MESSAGE_DEBUG: (6d11656c1e5f4982): Value "side/Read.out" materialized.
root: INFO: 2017-05-10T01:38:23.819Z: JOB_MESSAGE_BASIC: (6d11656c1e5f43f3): Executing operation compute/SideInput-s3
root: INFO: 2017-05-10T01:38:23.830Z: JOB_MESSAGE_DEBUG: (6d11656c1e5f46e1): Value "compute/SideInput-s3.output" materialized.
root: INFO: 2017-05-10T01:38:23.840Z: JOB_MESSAGE_BASIC: (558b513d87824a38): Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2017-05-10T01:38:24.044Z: JOB_MESSAGE_DEBUG: (284ca35a2d616197): Starting worker pool setup.
root: INFO: 2017-05-10T01:38:24.046Z: JOB_MESSAGE_BASIC: (284ca35a2d616d89): Starting 1 workers...
root: INFO: 2017-05-10T01:38:24.060Z: JOB_MESSAGE_DEBUG: (558b513d87824f3b): Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2017-05-10T01:38:24.069Z: JOB_MESSAGE_BASIC: (558b513d87824c93): Executing operation start/Read+compute/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T01:38:24.074Z: JOB_MESSAGE_BASIC: (6d11656c1e5f49cf): Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-05-10T01:40:52.405Z: JOB_MESSAGE_DETAILED: (f77e14fb3a3d9825): Workers have started successfully.
root: INFO: 2017-05-10T01:42:36.472Z: JOB_MESSAGE_ERROR: (cb0fcf6ab63672): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T01:42:39.858Z: JOB_MESSAGE_ERROR: (cb0fcf6ab63842): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T01:42:43.230Z: JOB_MESSAGE_ERROR: (cb0fcf6ab632f5): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T01:42:46.599Z: JOB_MESSAGE_ERROR: (cb0fcf6ab63da8): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T01:42:50.008Z: JOB_MESSAGE_ERROR: (cb0fcf6ab6385b): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T01:42:53.406Z: JOB_MESSAGE_ERROR: (cb0fcf6ab6330e): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 581, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 166, in execute
    op.start()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/native_operations.py", line 49, in start
    self._current_progress = get_progress()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/workercustomsources.py", line 126, in get_progress
    return iobase.ReaderProgress(
AttributeError: 'module' object has no attribute 'ReaderProgress'

root: INFO: 2017-05-10T01:42:53.551Z: JOB_MESSAGE_DEBUG: (558b513d878240ec): Executing failure step failure20
root: INFO: 2017-05-10T01:42:53.554Z: JOB_MESSAGE_ERROR: (558b513d87824042): Workflow failed. Causes: (558b513d87824240): S05:start/Read+compute/Do+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write failed.
root: INFO: 2017-05-10T01:42:53.623Z: JOB_MESSAGE_DETAILED: (28e4844b147a297a): Cleaning up.
root: INFO: 2017-05-10T01:42:53.735Z: JOB_MESSAGE_DEBUG: (28e4844b147a29bc): Starting worker pool teardown.
root: INFO: 2017-05-10T01:42:53.737Z: JOB_MESSAGE_BASIC: (28e4844b147a29fe): Stopping worker pool...
root: INFO: 2017-05-10T01:44:18.746Z: JOB_MESSAGE_BASIC: (28e4844b147a2261): Worker pool stopped.
root: INFO: 2017-05-10T01:44:18.788Z: JOB_MESSAGE_DEBUG: (28e4844b147a2c0e): Tearing down pending resources...
root: INFO: Job 2017-05-09_18_38_19-10569987656720048055 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 15 tests in 1466.598s

FAILED (errors=14)
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_20_22-7671973164766496890?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_26_03-17445787645204623552?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_32_14-16246056621494600472?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_38_19-10569987656720048055?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_20_22-4969442310096312321?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_26_21-2200428839861560191?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_33_24-13851441563056050815?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_20_22-8294999073061692102?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_26_16-11540265905883083825?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_32_17-7192357206309032277?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_38_33-1006407180017650573?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_20_21-14971386839334709645?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_25_46-15991440303187381446?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_31_49-14241720216081775588?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-09_18_37_34-15670525954546806359?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Not sending mail to unregistered user kirpichov@google.com
Not sending mail to unregistered user ccy@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2173

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2173/display/redirect?page=changes>

Changes:

[klk] Revise javadoc for sdk, state, options, annotations, window, (and misc)

[altay] Remove templates from wordcount example

[altay] Remove hadoop io readme

[tgroh] Shade JSR305 in the DirectRunner

------------------------------------------
[...truncated 579.48 KB...]
  File "/usr/lib/python2.7/pickle.py", line 600, in save_list
    self._batch_appends(iter(obj))
  File "/usr/lib/python2.7/pickle.py", line 633, in _batch_appends
    save(x)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 649, in save_dict
    self._batch_setitems(obj.iteritems())
  File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems
    save(v)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "/usr/lib/python2.7/pickle.py", line 600, in save_list
    self._batch_appends(iter(obj))
  File "/usr/lib/python2.7/pickle.py", line 633, in _batch_appends
    save(x)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 649, in save_dict
    self._batch_setitems(obj.iteritems())
  File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems
    save(v)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 649, in save_dict
    self._batch_setitems(obj.iteritems())
  File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems
    save(v)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "/usr/lib/python2.7/pickle.py", line 548, in save_tuple
    save(element)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 649, in save_dict
    self._batch_setitems(obj.iteritems())
  File "/usr/lib/python2.7/pickle.py", line 672, in _batch_setitems
    tmp.append(items.next())
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 156, in test_multi_valued_singleton_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 35, in run
    self.result = super(TestDataflowRunner, self).run(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 249, in run
    self.dataflow_client.create_job(self.job), self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 423, in create_job
    self.create_job_description(job)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 446, in create_job_description
    job.options, file_copy=self._gcs_file_copy)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 294, in stage_job_resources
    file_copy(setup_options.requirements_file, staged_path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 386, in _gcs_file_copy
    self.stage_file(to_folder, to_name, f)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 405, in stage_file
    response = self._storage_client.objects.Insert(request, upload=upload)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py",> line 971, in Insert
    download=download)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 721, in _RunMethod
    http, http_request, **opts)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 341, in MakeRequest
    check_response_func=check_response_func)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 391, in _MakeRequestNoRetry
    redirections=redirections, connection_type=connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/oauth2client/transport.py",> line 169, in new_request
    redirections, connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1609, in request
    (response, content) = self._request(conn, authority, uri, request_uri, method, body, headers, redirections, cachekey)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1351, in _request
    (response, content) = self._conn_request(conn, request_uri, method, body, headers)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1307, in _conn_request
    response = conn.getresponse()
  File "/usr/lib/python2.7/httplib.py", line 1089, in getresponse
    response.begin()
  File "/usr/lib/python2.7/httplib.py", line 444, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python2.7/httplib.py", line 400, in _read_status
    line = self.fp.readline(_MAXLINE + 1)
  File "/usr/lib/python2.7/socket.py", line 476, in readline
    data = self._sock.recv(self._rbufsize)
  File "/usr/lib/python2.7/ssl.py", line 341, in recv
    return self.read(buflen)
  File "/usr/lib/python2.7/ssl.py", line 260, in read
    return self._sslobj.read(len)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 235, in test_par_do_with_multiple_outputs_and_using_return
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)'

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 219, in test_par_do_with_multiple_outputs_and_using_yield
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest)'

======================================================================
ERROR: test_read_from_text_metrics (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 190, in test_read_from_text_metrics
    res = pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_read_from_text_metrics (apache_beam.transforms.ptransform_test.PTransformTest)'

----------------------------------------------------------------------
Ran 15 tests in 902.697s

FAILED (errors=15)
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Not sending mail to unregistered user kirpichov@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2172

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2172/display/redirect?page=changes>

Changes:

[tgroh] Make PCollectionViewWindow package private

------------------------------------------
[...truncated 957.57 KB...]
    self.save(obj)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 649, in save_dict
    self._batch_setitems(obj.iteritems())
  File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems
    save(v)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 396, in save_reduce
    save(cls)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 80, in wrapper
    containing_class_and_name = find_containing_class(obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 57, in find_containing_class
    return find_containing_class_inner(sys.modules[nested_class.__module__])
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 54, in find_containing_class_inner
    res = find_containing_class_inner(v)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 53, in find_containing_class_inner
    elif isinstance(v, (type, types.ClassType)) and hasattr(v, '__dict__'):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 177, in test_iterable_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 35, in run
    self.result = super(TestDataflowRunner, self).run(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 224, in run
    self.job = apiclient.Job(pipeline._options)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 311, in __init__
    self.google_cloud_options = options.view_as(GoogleCloudOptions)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/options/pipeline_options.py",> line 213, in view_as
    view = cls(self._flags)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/options/pipeline_options.py",> line 136, in __init__
    parser = BeamArgumentParser()
  File "/usr/lib/python2.7/argparse.py", line 1587, in __init__
    self._optionals = add_group(_('optional arguments'))
  File "/usr/lib/python2.7/gettext.py", line 581, in gettext
    return dgettext(_current_domain, message)
  File "/usr/lib/python2.7/gettext.py", line 545, in dgettext
    codeset=_localecodesets.get(domain))
  File "/usr/lib/python2.7/gettext.py", line 480, in translation
    mofiles = find(domain, localedir, languages, all=1)
  File "/usr/lib/python2.7/gettext.py", line 456, in find
    if os.path.exists(mofile):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/lib/python2.7/genericpath.py",> line 18, in exists
    os.stat(path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 156, in test_multi_valued_singleton_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 35, in run
    self.result = super(TestDataflowRunner, self).run(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 249, in run
    self.dataflow_client.create_job(self.job), self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 423, in create_job
    self.create_job_description(job)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 446, in create_job_description
    job.options, file_copy=self._gcs_file_copy)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 294, in stage_job_resources
    file_copy(setup_options.requirements_file, staged_path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 386, in _gcs_file_copy
    self.stage_file(to_folder, to_name, f)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 405, in stage_file
    response = self._storage_client.objects.Insert(request, upload=upload)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py",> line 971, in Insert
    download=download)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 721, in _RunMethod
    http, http_request, **opts)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 341, in MakeRequest
    check_response_func=check_response_func)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 391, in _MakeRequestNoRetry
    redirections=redirections, connection_type=connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/oauth2client/transport.py",> line 169, in new_request
    redirections, connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1609, in request
    (response, content) = self._request(conn, authority, uri, request_uri, method, body, headers, redirections, cachekey)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1351, in _request
    (response, content) = self._conn_request(conn, request_uri, method, body, headers)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1307, in _conn_request
    response = conn.getresponse()
  File "/usr/lib/python2.7/httplib.py", line 1089, in getresponse
    response.begin()
  File "/usr/lib/python2.7/httplib.py", line 444, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python2.7/httplib.py", line 400, in _read_status
    line = self.fp.readline(_MAXLINE + 1)
  File "/usr/lib/python2.7/socket.py", line 476, in readline
    data = self._sock.recv(self._rbufsize)
  File "/usr/lib/python2.7/ssl.py", line 341, in recv
    return self.read(buflen)
  File "/usr/lib/python2.7/ssl.py", line 260, in read
    return self._sslobj.read(len)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 291, in test_as_dict_twice
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 200, in test_as_list_and_as_dict_side_inputs
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 269, in test_as_list_twice
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest)'

----------------------------------------------------------------------
Ran 15 tests in 1541.468s

FAILED (errors=15)
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user kirpichov@google.com

Build failed in Jenkins: beam_PostCommit_Python_Verify #2171

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2171/display/redirect?page=changes>

Changes:

[dhalperi] Handle Errors in the DirectRunner

[dhalperi] Shade Dependencies of the DirectRunner

[altay] Remove old imports meant for backwards compatibility

[kirpichov] KafkaIO API clean up

[kirpichov] KafkaIO javadoc fixups

[dhalperi] [BEAM-2229] GcsFileSystem: handle empty files

------------------------------------------
[...truncated 572.26 KB...]
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 163, in new_save_module_dict
    return old_save_module_dict(pickler, obj)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 841, in save_module_dict
    StockPickler.save_dict(pickler, obj)
  File "/usr/lib/python2.7/pickle.py", line 649, in save_dict
    self._batch_setitems(obj.iteritems())
  File "/usr/lib/python2.7/pickle.py", line 681, in _batch_setitems
    save(v)
  File "/usr/lib/python2.7/pickle.py", line 331, in save
    self.save_reduce(obj=obj, *rv)
  File "/usr/lib/python2.7/pickle.py", line 419, in save_reduce
    save(state)
  File "/usr/lib/python2.7/pickle.py", line 286, in save
    f(self, obj) # Call unbound method with explicit self
  File "/usr/lib/python2.7/pickle.py", line 548, in save_tuple
    save(element)
  File "/usr/lib/python2.7/pickle.py", line 284, in save
    f = self.dispatch.get(t)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 177, in test_iterable_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 162, in run
    if test_runner_api and self._verify_runner_api_compatible():
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 332, in _verify_runner_api_compatible
    self.visit(Visitor())
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 196, in visit
    self._root_transform().visit(visitor, self, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 489, in visit
    part.visit(visitor, pipeline, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 489, in visit
    part.visit(visitor, pipeline, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 489, in visit
    part.visit(visitor, pipeline, visited)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 492, in visit
    visitor.visit_transform(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 324, in visit_transform
    enable_trace=False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/internal/pickler.py",> line 215, in loads
    return dill.loads(s)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 277, in loads
    return load(file)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/dill/dill.py",> line 266, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1170, in load_binput
    self.memo[repr(i)] = self.stack[-1]
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 156, in test_multi_valued_singleton_side_input
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 35, in run
    self.result = super(TestDataflowRunner, self).run(pipeline)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 249, in run
    self.dataflow_client.create_job(self.job), self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 423, in create_job
    self.create_job_description(job)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 446, in create_job_description
    job.options, file_copy=self._gcs_file_copy)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 294, in stage_job_resources
    file_copy(setup_options.requirements_file, staged_path)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py",> line 166, in wrapper
    return fun(*args, **kwargs)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 386, in _gcs_file_copy
    self.stage_file(to_folder, to_name, f)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 405, in stage_file
    response = self._storage_client.objects.Insert(request, upload=upload)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py",> line 971, in Insert
    download=download)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 721, in _RunMethod
    http, http_request, **opts)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 341, in MakeRequest
    check_response_func=check_response_func)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 391, in _MakeRequestNoRetry
    redirections=redirections, connection_type=connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/oauth2client/transport.py",> line 169, in new_request
    redirections, connection_type)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1609, in request
    (response, content) = self._request(conn, authority, uri, request_uri, method, body, headers, redirections, cachekey)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1351, in _request
    (response, content) = self._conn_request(conn, request_uri, method, body, headers)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1307, in _conn_request
    response = conn.getresponse()
  File "/usr/lib/python2.7/httplib.py", line 1089, in getresponse
    response.begin()
  File "/usr/lib/python2.7/httplib.py", line 444, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python2.7/httplib.py", line 400, in _read_status
    line = self.fp.readline(_MAXLINE + 1)
  File "/usr/lib/python2.7/socket.py", line 476, in readline
    data = self._sock.recv(self._rbufsize)
  File "/usr/lib/python2.7/ssl.py", line 341, in recv
    return self.read(buflen)
  File "/usr/lib/python2.7/ssl.py", line 260, in read
    return self._sslobj.read(len)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest)'

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 235, in test_par_do_with_multiple_outputs_and_using_return
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest)'

======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 219, in test_par_do_with_multiple_outputs_and_using_yield
    pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest)'

======================================================================
ERROR: test_read_from_text_metrics (apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 812, in run
    test(orig)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 45, in __call__
    return self.run(*arg, **kwarg)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 133, in run
    self.runTest(result)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/case.py",> line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
    testMethod()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/transforms/ptransform_test.py",> line 190, in test_read_from_text_metrics
    res = pipeline.run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/test_pipeline.py",> line 91, in run
    result = super(TestPipeline, self).run()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 164, in run
    self.to_runner_api(), self.runner, self._options).run(False)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/pipeline.py",> line 173, in run
    return self.runner.run(self)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 44, in run
    self.result.wait_until_finish()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 769, in wait_until_finish
    time.sleep(5.0)
  File "<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",> line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_read_from_text_metrics (apache_beam.transforms.ptransform_test.PTransformTest)'

----------------------------------------------------------------------
Ran 15 tests in 902.527s

FAILED (errors=15)
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user kirpichov@google.com