You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2017/01/23 23:35:36 UTC

Build failed in Jenkins: beam_PostCommit_Python_Verify #1075

See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1075/changes>

Changes:

[chamikara] Increments major version used by Dataflow runner to 5

[robertwb] Remove dataflow_test.py

[robertwb] Code cleanup now that all runners support windowed side inputs.

------------------------------------------
[...truncated 8425 lines...]
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert:even/UnKey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s15"
        }, 
        "serialized_fn": "<string of 1052 bytes>", 
        "user_name": "assert:even/UnKey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s17", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert:even/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s16"
        }, 
        "serialized_fn": "<string of 1212 bytes>", 
        "user_name": "assert:even/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 id: u'2017-01-23_15_30_54-12011182360156094633'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-01-23_15_30_54-12011182360156094633]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-01-23_15_30_54-12011182360156094633
root: INFO: Job 2017-01-23_15_30_54-12011182360156094633 is in state JOB_STATE_RUNNING
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab092b: 2017-01-23T23:30:55.403Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312046): Checking required Cloud APIs are enabled.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b0b: 2017-01-23T23:30:55.883Z: JOB_MESSAGE_DEBUG: (3fadbfc3db312195): Combiner lifting skipped for step assert_that/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b0d: 2017-01-23T23:30:55.885Z: JOB_MESSAGE_DEBUG: (3fadbfc3db312ac3): Combiner lifting skipped for step assert:even/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b0f: 2017-01-23T23:30:55.887Z: JOB_MESSAGE_DEBUG: (3fadbfc3db3123f1): Combiner lifting skipped for step assert:odd/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b12: 2017-01-23T23:30:55.890Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312d1f): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b14: 2017-01-23T23:30:55.892Z: JOB_MESSAGE_DETAILED: (3fadbfc3db31264d): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b1e: 2017-01-23T23:30:55.902Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312b05): Annotating graph with Autotuner information.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b29: 2017-01-23T23:30:55.913Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3129f8): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b2f: 2017-01-23T23:30:55.919Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312c54): Fusing consumer assert:odd/ToVoidKey into assert:odd/WindowInto
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b32: 2017-01-23T23:30:55.922Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312582): Fusing consumer assert:odd/UnKey into assert:odd/Group/GroupByWindow
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b34: 2017-01-23T23:30:55.924Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312eb0): Fusing consumer assert:even/UnKey into assert:even/Group/GroupByWindow
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b37: 2017-01-23T23:30:55.927Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3127de): Fusing consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b39: 2017-01-23T23:30:55.929Z: JOB_MESSAGE_DETAILED: (3fadbfc3db31210c): Fusing consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b3b: 2017-01-23T23:30:55.931Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312a3a): Fusing consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b3d: 2017-01-23T23:30:55.933Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312368): Fusing consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b3f: 2017-01-23T23:30:55.935Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312c96): Fusing consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b42: 2017-01-23T23:30:55.938Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3125c4): Fusing consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b44: 2017-01-23T23:30:55.940Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312ef2): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b46: 2017-01-23T23:30:55.942Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312820): Fusing consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b48: 2017-01-23T23:30:55.944Z: JOB_MESSAGE_DETAILED: (3fadbfc3db31214e): Fusing consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b4a: 2017-01-23T23:30:55.946Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312a7c): Fusing consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b4c: 2017-01-23T23:30:55.948Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3123aa): Fusing consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b4f: 2017-01-23T23:30:55.951Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312cd8): Fusing consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b51: 2017-01-23T23:30:55.953Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312606): Fusing consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b53: 2017-01-23T23:30:55.955Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312f34): Fusing consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b55: 2017-01-23T23:30:55.957Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312862): Fusing consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b57: 2017-01-23T23:30:55.959Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312190): Fusing consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b59: 2017-01-23T23:30:55.961Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312abe): Fusing consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b5b: 2017-01-23T23:30:55.963Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3123ec): Fusing consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0bb1: 2017-01-23T23:30:56.049Z: JOB_MESSAGE_DEBUG: (3fadbfc3db3120c0): Workflow config is missing a default resource spec.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0bb3: 2017-01-23T23:30:56.051Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3129ee): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0bf6: 2017-01-23T23:30:56.118Z: JOB_MESSAGE_DEBUG: (e31afff0dc916c6a): Adding workflow start and stop steps.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0c28: 2017-01-23T23:30:56.168Z: JOB_MESSAGE_DEBUG: (dd9172f79e564817): Assigning stage ids.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0ccc: 2017-01-23T23:30:56.332Z: JOB_MESSAGE_DEBUG: (aaac8d07a592185a): Executing wait step start2
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cd8: 2017-01-23T23:30:56.344Z: JOB_MESSAGE_DEBUG: (e5b879d3fc421531): Executing operation Some Numbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0ce6: 2017-01-23T23:30:56.358Z: JOB_MESSAGE_DEBUG: (e31afff0dc916fc1): Value "Some Numbers.out" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cf1: 2017-01-23T23:30:56.369Z: JOB_MESSAGE_BASIC: S01: (aaac8d07a5921fae): Executing operation assert:odd/Group/Create
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cf4: 2017-01-23T23:30:56.372Z: JOB_MESSAGE_BASIC: S02: (3fadbfc3db3128e1): Executing operation assert:even/Group/Create
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cf5: 2017-01-23T23:30:56.373Z: JOB_MESSAGE_BASIC: S03: (6e37547a5e2c658a): Executing operation assert_that/Group/Create
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dbe: 2017-01-23T23:30:56.574Z: JOB_MESSAGE_DEBUG: (e9a91083e1434023): Starting worker pool setup.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dc0: 2017-01-23T23:30:56.576Z: JOB_MESSAGE_BASIC: (e9a91083e14348bd): Starting 1 workers...
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dd2: 2017-01-23T23:30:56.594Z: JOB_MESSAGE_DEBUG: (708a4aa4ce81981a): Value "assert:odd/Group/Session" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dd4: 2017-01-23T23:30:56.596Z: JOB_MESSAGE_DEBUG: (dd9172f79e564c8e): Value "assert_that/Group/Session" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dea: 2017-01-23T23:30:56.618Z: JOB_MESSAGE_DEBUG: (3bd4ccbdb25e4eac): Value "assert:even/Group/Session" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0df5: 2017-01-23T23:30:56.629Z: JOB_MESSAGE_BASIC: S04: (e5b879d3fc4213da): Executing operation ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdabed2f: 2017-01-23T23:31:53.775Z: JOB_MESSAGE_DETAILED: (d3212f834d43fd2a): Workers have started successfully.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd6e4: 2017-01-23T23:33:59.140Z: JOB_MESSAGE_ERROR: (d3f0897736e1c274): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd7cb: 2017-01-23T23:33:59.371Z: JOB_MESSAGE_ERROR: (d3f0897736e1c8b9): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd882: 2017-01-23T23:33:59.554Z: JOB_MESSAGE_ERROR: (d3f0897736e1c3b3): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd965: 2017-01-23T23:33:59.781Z: JOB_MESSAGE_ERROR: (d3f0897736e1c9f8): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd98f: 2017-01-23T23:33:59.823Z: JOB_MESSAGE_DEBUG: (3fadbfc3db312ff5): Executing failure step failure1
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd991: 2017-01-23T23:33:59.825Z: JOB_MESSAGE_ERROR: (3fadbfc3db312923): Workflow failed. Causes: (e5b879d3fc42115d): S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write failed.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd9cf: 2017-01-23T23:33:59.887Z: JOB_MESSAGE_DETAILED: (580466590c7e466a): Cleaning up.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadda4c: 2017-01-23T23:34:00.012Z: JOB_MESSAGE_DEBUG: (580466590c7e4304): Starting worker pool teardown.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadda4e: 2017-01-23T23:34:00.014Z: JOB_MESSAGE_BASIC: (580466590c7e4f9e): Stopping worker pool...
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdaf12d6: 2017-01-23T23:35:20.022Z: JOB_MESSAGE_BASIC: (580466590c7e4285): Worker pool stopped.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdaf16d8: 2017-01-23T23:35:21.048Z: JOB_MESSAGE_DEBUG: (580466590c7e4853): Tearing down pending resources...
root: INFO: Job 2017-01-23_15_30_54-12011182360156094633 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1186.872s

FAILED (errors=4)
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: beam_PostCommit_Python_Verify #1078

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1078/>

------------------------------------------
[...truncated 8270 lines...]
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert:even/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s16"
        }, 
        "serialized_fn": "<string of 1212 bytes>", 
        "user_name": "assert:even/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 id: u'2017-01-24_07_18_48-4572845589957073573'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-01-24_07_18_48-4572845589957073573]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-01-24_07_18_48-4572845589957073573
root: INFO: Job 2017-01-24_07_18_48-4572845589957073573 is in state JOB_STATE_RUNNING
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee314: 2017-01-24T15:18:50.900Z: JOB_MESSAGE_DETAILED: (70e8f858d26e09d5): Checking required Cloud APIs are enabled.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee915: 2017-01-24T15:18:52.437Z: JOB_MESSAGE_DEBUG: (70e8f858d26e06c2): Combiner lifting skipped for step assert_that/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee917: 2017-01-24T15:18:52.439Z: JOB_MESSAGE_DEBUG: (70e8f858d26e044c): Combiner lifting skipped for step assert:even/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee91a: 2017-01-24T15:18:52.442Z: JOB_MESSAGE_DEBUG: (70e8f858d26e01d6): Combiner lifting skipped for step assert:odd/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee927: 2017-01-24T15:18:52.455Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0f60): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee93c: 2017-01-24T15:18:52.476Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0cea): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee957: 2017-01-24T15:18:52.503Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0312): Annotating graph with Autotuner information.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee989: 2017-01-24T15:18:52.553Z: JOB_MESSAGE_DETAILED: (70e8f858d26e04eb): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee98f: 2017-01-24T15:18:52.559Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0fff): Fusing consumer assert:odd/ToVoidKey into assert:odd/WindowInto
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee992: 2017-01-24T15:18:52.562Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0d89): Fusing consumer assert:odd/UnKey into assert:odd/Group/GroupByWindow
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee995: 2017-01-24T15:18:52.565Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0b13): Fusing consumer assert:even/UnKey into assert:even/Group/GroupByWindow
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee998: 2017-01-24T15:18:52.568Z: JOB_MESSAGE_DETAILED: (70e8f858d26e089d): Fusing consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee99a: 2017-01-24T15:18:52.570Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0627): Fusing consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee99d: 2017-01-24T15:18:52.573Z: JOB_MESSAGE_DETAILED: (70e8f858d26e03b1): Fusing consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a0: 2017-01-24T15:18:52.576Z: JOB_MESSAGE_DETAILED: (70e8f858d26e013b): Fusing consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a2: 2017-01-24T15:18:52.578Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0ec5): Fusing consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a4: 2017-01-24T15:18:52.580Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0c4f): Fusing consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a7: 2017-01-24T15:18:52.583Z: JOB_MESSAGE_DETAILED: (70e8f858d26e09d9): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a9: 2017-01-24T15:18:52.585Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0763): Fusing consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9ab: 2017-01-24T15:18:52.587Z: JOB_MESSAGE_DETAILED: (70e8f858d26e04ed): Fusing consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9ad: 2017-01-24T15:18:52.589Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0277): Fusing consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9af: 2017-01-24T15:18:52.591Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0001): Fusing consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b1: 2017-01-24T15:18:52.593Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0d8b): Fusing consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b3: 2017-01-24T15:18:52.595Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0b15): Fusing consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b6: 2017-01-24T15:18:52.598Z: JOB_MESSAGE_DETAILED: (70e8f858d26e089f): Fusing consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b9: 2017-01-24T15:18:52.601Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0629): Fusing consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9bc: 2017-01-24T15:18:52.604Z: JOB_MESSAGE_DETAILED: (70e8f858d26e03b3): Fusing consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9be: 2017-01-24T15:18:52.606Z: JOB_MESSAGE_DETAILED: (70e8f858d26e013d): Fusing consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9c0: 2017-01-24T15:18:52.608Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0ec7): Fusing consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eea6c: 2017-01-24T15:18:52.780Z: JOB_MESSAGE_DEBUG: (70e8f858d26e0143): Workflow config is missing a default resource spec.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eea7d: 2017-01-24T15:18:52.797Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0ecd): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eead2: 2017-01-24T15:18:52.882Z: JOB_MESSAGE_DEBUG: (a0290f97a19667a2): Adding workflow start and stop steps.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeb04: 2017-01-24T15:18:52.932Z: JOB_MESSAGE_DEBUG: (c297fa4537c49501): Assigning stage ids.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeb94: 2017-01-24T15:18:53.076Z: JOB_MESSAGE_DEBUG: (70e8f858d26e09e1): Executing wait step start2
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeba6: 2017-01-24T15:18:53.094Z: JOB_MESSAGE_DEBUG: (65a3ce102518fe29): Executing operation Some Numbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebb3: 2017-01-24T15:18:53.107Z: JOB_MESSAGE_DEBUG: (2eb55d8630d2cd6a): Value "Some Numbers.out" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebc1: 2017-01-24T15:18:53.121Z: JOB_MESSAGE_BASIC: S01: (2eb55d8630d2c79b): Executing operation assert:odd/Group/Create
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebc4: 2017-01-24T15:18:53.124Z: JOB_MESSAGE_BASIC: S03: (9a4f486fa95aca6f): Executing operation assert_that/Group/Create
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebd1: 2017-01-24T15:18:53.137Z: JOB_MESSAGE_BASIC: S02: (2ab7216f4602c8a3): Executing operation assert:even/Group/Create
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eec8e: 2017-01-24T15:18:53.326Z: JOB_MESSAGE_DEBUG: (35722c8b1092de91): Starting worker pool setup.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eec90: 2017-01-24T15:18:53.328Z: JOB_MESSAGE_BASIC: (35722c8b1092d193): Starting 1 workers...
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeca5: 2017-01-24T15:18:53.349Z: JOB_MESSAGE_DEBUG: (70e8f858d26e04f5): Value "assert:odd/Group/Session" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeca7: 2017-01-24T15:18:53.351Z: JOB_MESSAGE_DEBUG: (5ca6d1bc2bc714ff): Value "assert:even/Group/Session" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeca9: 2017-01-24T15:18:53.353Z: JOB_MESSAGE_DEBUG: (940ba07495593a6c): Value "assert_that/Group/Session" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eecb7: 2017-01-24T15:18:53.367Z: JOB_MESSAGE_BASIC: S04: (5ca6d1bc2bc71139): Executing operation ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d11092c8: 2017-01-24T15:20:41.416Z: JOB_MESSAGE_DETAILED: (45b4fae73847d61d): Workers have started successfully.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1114f4b: 2017-01-24T15:21:29.675Z: JOB_MESSAGE_ERROR: (51b0ae8abdf9362e): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d111504a: 2017-01-24T15:21:29.930Z: JOB_MESSAGE_ERROR: (51b0ae8abdf934a5): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d11150f7: 2017-01-24T15:21:30.103Z: JOB_MESSAGE_ERROR: (51b0ae8abdf9331c): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d111517f: 2017-01-24T15:21:30.239Z: JOB_MESSAGE_ERROR: (51b0ae8abdf936da): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115228: 2017-01-24T15:21:30.408Z: JOB_MESSAGE_ERROR: (51b0ae8abdf93551): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115254: 2017-01-24T15:21:30.452Z: JOB_MESSAGE_DEBUG: (6c3718a70c324aa): Executing failure step failure1
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115258: 2017-01-24T15:21:30.456Z: JOB_MESSAGE_ERROR: (6c3718a70c32f94): Workflow failed. Causes: (5ca6d1bc2bc7190c): S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write failed.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115299: 2017-01-24T15:21:30.521Z: JOB_MESSAGE_DETAILED: (66844a3ff238a206): Cleaning up.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115313: 2017-01-24T15:21:30.643Z: JOB_MESSAGE_DEBUG: (66844a3ff238a254): Starting worker pool teardown.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115315: 2017-01-24T15:21:30.645Z: JOB_MESSAGE_BASIC: (66844a3ff238a2a2): Stopping worker pool...
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d112648c: 2017-01-24T15:22:40.652Z: JOB_MESSAGE_BASIC: (66844a3ff238ab17): Worker pool stopped.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1126893: 2017-01-24T15:22:41.683Z: JOB_MESSAGE_DEBUG: (66844a3ff238ac01): Tearing down pending resources...
root: INFO: Job 2017-01-24_07_18_48-4572845589957073573 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1024.260s

FAILED (errors=4)
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: beam_PostCommit_Python_Verify #1077

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1077/>

------------------------------------------
[...truncated 8519 lines...]
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa8b: 2017-01-24T09:18:30.027Z: JOB_MESSAGE_DETAILED: (87856bc85f4e273e): Fusing consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa8d: 2017-01-24T09:18:30.029Z: JOB_MESSAGE_DETAILED: (87856bc85f4e24e0): Fusing consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa8f: 2017-01-24T09:18:30.031Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2282): Fusing consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa91: 2017-01-24T09:18:30.033Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2024): Fusing consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa93: 2017-01-24T09:18:30.035Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2dc6): Fusing consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa95: 2017-01-24T09:18:30.037Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2b68): Fusing consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa97: 2017-01-24T09:18:30.039Z: JOB_MESSAGE_DETAILED: (87856bc85f4e290a): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa99: 2017-01-24T09:18:30.041Z: JOB_MESSAGE_DETAILED: (87856bc85f4e26ac): Fusing consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa9b: 2017-01-24T09:18:30.043Z: JOB_MESSAGE_DETAILED: (87856bc85f4e244e): Fusing consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa9d: 2017-01-24T09:18:30.045Z: JOB_MESSAGE_DETAILED: (87856bc85f4e21f0): Fusing consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa9f: 2017-01-24T09:18:30.047Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2f92): Fusing consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa1: 2017-01-24T09:18:30.049Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2d34): Fusing consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa3: 2017-01-24T09:18:30.051Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2ad6): Fusing consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa5: 2017-01-24T09:18:30.053Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2878): Fusing consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa7: 2017-01-24T09:18:30.055Z: JOB_MESSAGE_DETAILED: (87856bc85f4e261a): Fusing consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa9: 2017-01-24T09:18:30.057Z: JOB_MESSAGE_DETAILED: (87856bc85f4e23bc): Fusing consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faab: 2017-01-24T09:18:30.059Z: JOB_MESSAGE_DETAILED: (87856bc85f4e215e): Fusing consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faad: 2017-01-24T09:18:30.061Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2f00): Fusing consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb00: 2017-01-24T09:18:30.144Z: JOB_MESSAGE_DEBUG: (87856bc85f4e250c): Workflow config is missing a default resource spec.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb03: 2017-01-24T09:18:30.147Z: JOB_MESSAGE_DETAILED: (87856bc85f4e22ae): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb21: 2017-01-24T09:18:30.177Z: JOB_MESSAGE_DEBUG: (31cd4de23232ef81): Adding workflow start and stop steps.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb53: 2017-01-24T09:18:30.227Z: JOB_MESSAGE_DEBUG: (f93568740c05d1b1): Assigning stage ids.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fba9: 2017-01-24T09:18:30.313Z: JOB_MESSAGE_DEBUG: (f93568740c05d098): Executing wait step start2
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbb9: 2017-01-24T09:18:30.329Z: JOB_MESSAGE_DEBUG: (ac2138444fc23d7d): Executing operation Some Numbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbee: 2017-01-24T09:18:30.382Z: JOB_MESSAGE_DEBUG: (a5eb5b5ec15467df): Value "Some Numbers.out" materialized.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbf7: 2017-01-24T09:18:30.391Z: JOB_MESSAGE_BASIC: S02: (f93568740c05d91c): Executing operation assert:even/Group/Create
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbfa: 2017-01-24T09:18:30.394Z: JOB_MESSAGE_BASIC: S03: (d90ef3f76c48fa2a): Executing operation assert_that/Group/Create
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fc1b: 2017-01-24T09:18:30.427Z: JOB_MESSAGE_BASIC: S01: (af4267f01a05b233): Executing operation assert:odd/Group/Create
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcc3: 2017-01-24T09:18:30.595Z: JOB_MESSAGE_DEBUG: (b9d847bf036a3ee3): Starting worker pool setup.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcc5: 2017-01-24T09:18:30.597Z: JOB_MESSAGE_BASIC: (b9d847bf036a36e5): Starting 1 workers...
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcd6: 2017-01-24T09:18:30.614Z: JOB_MESSAGE_DEBUG: (f1cc619595f671fd): Value "assert_that/Group/Session" materialized.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcd8: 2017-01-24T09:18:30.616Z: JOB_MESSAGE_DEBUG: (f93568740c05d2a8): Value "assert:even/Group/Session" materialized.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fce3: 2017-01-24T09:18:30.627Z: JOB_MESSAGE_BASIC: S04: (87856bc85f4e2cc3): Executing operation ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc6b696: 2017-01-24T09:20:23.702Z: JOB_MESSAGE_DETAILED: (43b1b18f7b1c6b33): Workers have started successfully.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7ca91: 2017-01-24T09:21:34.353Z: JOB_MESSAGE_ERROR: (72f067c206330db1): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cb72: 2017-01-24T09:21:34.578Z: JOB_MESSAGE_ERROR: (72f067c2063301ae): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cc0f: 2017-01-24T09:21:34.735Z: JOB_MESSAGE_ERROR: (72f067c206330e98): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7ccaf: 2017-01-24T09:21:34.895Z: JOB_MESSAGE_ERROR: (72f067c206330b82): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cd46: 2017-01-24T09:21:35.046Z: JOB_MESSAGE_ERROR: (72f067c20633086c): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7ce14: 2017-01-24T09:21:35.252Z: JOB_MESSAGE_ERROR: (72f067c206330c69): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cedc: 2017-01-24T09:21:35.452Z: JOB_MESSAGE_ERROR: (72f067c206330066): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cfad: 2017-01-24T09:21:35.661Z: JOB_MESSAGE_ERROR: (72f067c206330463): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cfd6: 2017-01-24T09:21:35.702Z: JOB_MESSAGE_DEBUG: (5129b9404837122d): Executing failure step failure1
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cfd9: 2017-01-24T09:21:35.705Z: JOB_MESSAGE_ERROR: (5129b94048371343): Workflow failed. Causes: (87856bc85f4e2d60): S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write failed.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7d016: 2017-01-24T09:21:35.766Z: JOB_MESSAGE_DETAILED: (90c2002d0fc71cca): Cleaning up.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7d094: 2017-01-24T09:21:35.892Z: JOB_MESSAGE_DEBUG: (90c2002d0fc71e84): Starting worker pool teardown.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7d096: 2017-01-24T09:21:35.894Z: JOB_MESSAGE_BASIC: (90c2002d0fc7103e): Stopping worker pool...
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc8f596: 2017-01-24T09:22:50.902Z: JOB_MESSAGE_BASIC: (90c2002d0fc71ad5): Worker pool stopped.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc8f99a: 2017-01-24T09:22:51.930Z: JOB_MESSAGE_DEBUG: (90c2002d0fc71003): Tearing down pending resources...
root: INFO: Job 2017-01-24_01_18_28-12399986343756263108 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1040.088s

FAILED (errors=4)
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: beam_PostCommit_Python_Verify #1076

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1076/>

------------------------------------------
[...truncated 8345 lines...]
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert:even/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s16"
        }, 
        "serialized_fn": "<string of 1212 bytes>", 
        "user_name": "assert:even/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 id: u'2017-01-23_19_18_55-6733368171544728562'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-01-23_19_18_55-6733368171544728562]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-01-23_19_18_55-6733368171544728562
root: INFO: Job 2017-01-23_19_18_55-6733368171544728562 is in state JOB_STATE_RUNNING
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bceef: 2017-01-24T03:18:57.519Z: JOB_MESSAGE_DETAILED: (e11bab53777fc3a3): Checking required Cloud APIs are enabled.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd4b2: 2017-01-24T03:18:58.994Z: JOB_MESSAGE_DEBUG: (e11bab53777fc1b2): Combiner lifting skipped for step assert_that/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd4b4: 2017-01-24T03:18:58.996Z: JOB_MESSAGE_DEBUG: (e11bab53777fc860): Combiner lifting skipped for step assert:even/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd4b6: 2017-01-24T03:18:58.998Z: JOB_MESSAGE_DEBUG: (e11bab53777fcf0e): Combiner lifting skipped for step assert:odd/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd4c0: 2017-01-24T03:18:59.008Z: JOB_MESSAGE_DETAILED: (e11bab53777fc5bc): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd4d0: 2017-01-24T03:18:59.024Z: JOB_MESSAGE_DETAILED: (e11bab53777fcc6a): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd4eb: 2017-01-24T03:18:59.051Z: JOB_MESSAGE_DETAILED: (e11bab53777fc722): Annotating graph with Autotuner information.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd51f: 2017-01-24T03:18:59.103Z: JOB_MESSAGE_DETAILED: (e11bab53777fc7d5): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd524: 2017-01-24T03:18:59.108Z: JOB_MESSAGE_DETAILED: (e11bab53777fc531): Fusing consumer assert:odd/ToVoidKey into assert:odd/WindowInto
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd527: 2017-01-24T03:18:59.111Z: JOB_MESSAGE_DETAILED: (e11bab53777fcbdf): Fusing consumer assert:odd/UnKey into assert:odd/Group/GroupByWindow
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd529: 2017-01-24T03:18:59.113Z: JOB_MESSAGE_DETAILED: (e11bab53777fc28d): Fusing consumer assert:even/UnKey into assert:even/Group/GroupByWindow
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd52b: 2017-01-24T03:18:59.115Z: JOB_MESSAGE_DETAILED: (e11bab53777fc93b): Fusing consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd52e: 2017-01-24T03:18:59.118Z: JOB_MESSAGE_DETAILED: (e11bab53777fcfe9): Fusing consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd530: 2017-01-24T03:18:59.120Z: JOB_MESSAGE_DETAILED: (e11bab53777fc697): Fusing consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd533: 2017-01-24T03:18:59.123Z: JOB_MESSAGE_DETAILED: (e11bab53777fcd45): Fusing consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd535: 2017-01-24T03:18:59.125Z: JOB_MESSAGE_DETAILED: (e11bab53777fc3f3): Fusing consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd538: 2017-01-24T03:18:59.128Z: JOB_MESSAGE_DETAILED: (e11bab53777fcaa1): Fusing consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd53a: 2017-01-24T03:18:59.130Z: JOB_MESSAGE_DETAILED: (e11bab53777fc14f): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd53d: 2017-01-24T03:18:59.133Z: JOB_MESSAGE_DETAILED: (e11bab53777fc7fd): Fusing consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd53f: 2017-01-24T03:18:59.135Z: JOB_MESSAGE_DETAILED: (e11bab53777fceab): Fusing consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd542: 2017-01-24T03:18:59.138Z: JOB_MESSAGE_DETAILED: (e11bab53777fc559): Fusing consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd544: 2017-01-24T03:18:59.140Z: JOB_MESSAGE_DETAILED: (e11bab53777fcc07): Fusing consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd547: 2017-01-24T03:18:59.143Z: JOB_MESSAGE_DETAILED: (e11bab53777fc2b5): Fusing consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd549: 2017-01-24T03:18:59.145Z: JOB_MESSAGE_DETAILED: (e11bab53777fc963): Fusing consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd54b: 2017-01-24T03:18:59.147Z: JOB_MESSAGE_DETAILED: (e11bab53777fc011): Fusing consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd54e: 2017-01-24T03:18:59.150Z: JOB_MESSAGE_DETAILED: (e11bab53777fc6bf): Fusing consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd551: 2017-01-24T03:18:59.153Z: JOB_MESSAGE_DETAILED: (e11bab53777fcd6d): Fusing consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd553: 2017-01-24T03:18:59.155Z: JOB_MESSAGE_DETAILED: (e11bab53777fc41b): Fusing consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd555: 2017-01-24T03:18:59.157Z: JOB_MESSAGE_DETAILED: (e11bab53777fcac9): Fusing consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd5ea: 2017-01-24T03:18:59.306Z: JOB_MESSAGE_DEBUG: (e11bab53777fc89d): Workflow config is missing a default resource spec.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd5f9: 2017-01-24T03:18:59.321Z: JOB_MESSAGE_DETAILED: (e11bab53777fcf4b): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd649: 2017-01-24T03:18:59.401Z: JOB_MESSAGE_DEBUG: (969c07ebca9e3abd): Adding workflow start and stop steps.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd68b: 2017-01-24T03:18:59.467Z: JOB_MESSAGE_DEBUG: (5f455b8a2539c32a): Assigning stage ids.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd6ea: 2017-01-24T03:18:59.562Z: JOB_MESSAGE_DEBUG: (5f455b8a2539ce5b): Executing wait step start2
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd6f5: 2017-01-24T03:18:59.573Z: JOB_MESSAGE_DEBUG: (5f455b8a2539c74f): Executing operation Some Numbers
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd701: 2017-01-24T03:18:59.585Z: JOB_MESSAGE_DEBUG: (306d5741977075a): Value "Some Numbers.out" materialized.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd70b: 2017-01-24T03:18:59.595Z: JOB_MESSAGE_BASIC: S01: (9e50051901c7956e): Executing operation assert:odd/Group/Create
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd70d: 2017-01-24T03:18:59.597Z: JOB_MESSAGE_BASIC: S03: (1d63a3bedcd2bbda): Executing operation assert_that/Group/Create
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd714: 2017-01-24T03:18:59.604Z: JOB_MESSAGE_BASIC: S02: (969c07ebca9e3a81): Executing operation assert:even/Group/Create
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd7d7: 2017-01-24T03:18:59.799Z: JOB_MESSAGE_DEBUG: (6d98c9c703d1e811): Starting worker pool setup.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd7d9: 2017-01-24T03:18:59.801Z: JOB_MESSAGE_BASIC: (6d98c9c703d1e0d7): Starting 1 workers...
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd7ed: 2017-01-24T03:18:59.821Z: JOB_MESSAGE_DEBUG: (237b0974c1c0d60e): Value "assert:odd/Group/Session" materialized.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd811: 2017-01-24T03:18:59.857Z: JOB_MESSAGE_DEBUG: (20a78c1d17658ea5): Value "assert_that/Group/Session" materialized.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd813: 2017-01-24T03:18:59.859Z: JOB_MESSAGE_DEBUG: (3450a6e186be988b): Value "assert:even/Group/Session" materialized.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7bd81f: 2017-01-24T03:18:59.871Z: JOB_MESSAGE_BASIC: S04: (20a78c1d17658e73): Executing operation ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7c9807: 2017-01-24T03:19:48.999Z: JOB_MESSAGE_DETAILED: (61f94197025dddb8): Workers have started successfully.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e41a7: 2017-01-24T03:21:37.959Z: JOB_MESSAGE_ERROR: (65c712345d54e3cf): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e427b: 2017-01-24T03:21:38.171Z: JOB_MESSAGE_ERROR: (65c712345d54e8a4): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e4304: 2017-01-24T03:21:38.308Z: JOB_MESSAGE_ERROR: (65c712345d54e3be): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e437c: 2017-01-24T03:21:38.428Z: JOB_MESSAGE_ERROR: (65c712345d54eed8): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e4402: 2017-01-24T03:21:38.562Z: JOB_MESSAGE_ERROR: (65c712345d54e9f2): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e44b7: 2017-01-24T03:21:38.743Z: JOB_MESSAGE_ERROR: (65c712345d54eec7): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e44e4: 2017-01-24T03:21:38.788Z: JOB_MESSAGE_DEBUG: (66893aba8d70245e): Executing failure step failure1
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e44e7: 2017-01-24T03:21:38.791Z: JOB_MESSAGE_ERROR: (66893aba8d702dd0): Workflow failed. Causes: (20a78c1d1765835c): S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write failed.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e4525: 2017-01-24T03:21:38.853Z: JOB_MESSAGE_DETAILED: (3450a6e186be9e26): Cleaning up.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e45a2: 2017-01-24T03:21:38.978Z: JOB_MESSAGE_DEBUG: (3450a6e186be9fc0): Starting worker pool teardown.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7e45a4: 2017-01-24T03:21:38.980Z: JOB_MESSAGE_BASIC: (3450a6e186be915a): Stopping worker pool...
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7f7e2c: 2017-01-24T03:22:58.988Z: JOB_MESSAGE_BASIC: (3450a6e186be93c1): Worker pool stopped.
root: INFO: 2017-01-23_19_18_55-6733368171544728562_00000159ce7f8239: 2017-01-24T03:23:00.025Z: JOB_MESSAGE_DEBUG: (3450a6e186be988f): Tearing down pending resources...
root: INFO: Job 2017-01-23_19_18_55-6733368171544728562 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1050.523s

FAILED (errors=4)
Build step 'Execute shell' marked build as failure