You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2017/01/24 15:22:50 UTC

Build failed in Jenkins: beam_PostCommit_Python_Verify #1078

See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1078/>

------------------------------------------
[...truncated 8270 lines...]
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert:even/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s16"
        }, 
        "serialized_fn": "<string of 1212 bytes>", 
        "user_name": "assert:even/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 id: u'2017-01-24_07_18_48-4572845589957073573'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-01-24_07_18_48-4572845589957073573]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-01-24_07_18_48-4572845589957073573
root: INFO: Job 2017-01-24_07_18_48-4572845589957073573 is in state JOB_STATE_RUNNING
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee314: 2017-01-24T15:18:50.900Z: JOB_MESSAGE_DETAILED: (70e8f858d26e09d5): Checking required Cloud APIs are enabled.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee915: 2017-01-24T15:18:52.437Z: JOB_MESSAGE_DEBUG: (70e8f858d26e06c2): Combiner lifting skipped for step assert_that/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee917: 2017-01-24T15:18:52.439Z: JOB_MESSAGE_DEBUG: (70e8f858d26e044c): Combiner lifting skipped for step assert:even/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee91a: 2017-01-24T15:18:52.442Z: JOB_MESSAGE_DEBUG: (70e8f858d26e01d6): Combiner lifting skipped for step assert:odd/Group: GroupByKey not followed by a combiner.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee927: 2017-01-24T15:18:52.455Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0f60): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee93c: 2017-01-24T15:18:52.476Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0cea): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee957: 2017-01-24T15:18:52.503Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0312): Annotating graph with Autotuner information.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee989: 2017-01-24T15:18:52.553Z: JOB_MESSAGE_DETAILED: (70e8f858d26e04eb): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee98f: 2017-01-24T15:18:52.559Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0fff): Fusing consumer assert:odd/ToVoidKey into assert:odd/WindowInto
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee992: 2017-01-24T15:18:52.562Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0d89): Fusing consumer assert:odd/UnKey into assert:odd/Group/GroupByWindow
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee995: 2017-01-24T15:18:52.565Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0b13): Fusing consumer assert:even/UnKey into assert:even/Group/GroupByWindow
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee998: 2017-01-24T15:18:52.568Z: JOB_MESSAGE_DETAILED: (70e8f858d26e089d): Fusing consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee99a: 2017-01-24T15:18:52.570Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0627): Fusing consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee99d: 2017-01-24T15:18:52.573Z: JOB_MESSAGE_DETAILED: (70e8f858d26e03b1): Fusing consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a0: 2017-01-24T15:18:52.576Z: JOB_MESSAGE_DETAILED: (70e8f858d26e013b): Fusing consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a2: 2017-01-24T15:18:52.578Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0ec5): Fusing consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a4: 2017-01-24T15:18:52.580Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0c4f): Fusing consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a7: 2017-01-24T15:18:52.583Z: JOB_MESSAGE_DETAILED: (70e8f858d26e09d9): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9a9: 2017-01-24T15:18:52.585Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0763): Fusing consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9ab: 2017-01-24T15:18:52.587Z: JOB_MESSAGE_DETAILED: (70e8f858d26e04ed): Fusing consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9ad: 2017-01-24T15:18:52.589Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0277): Fusing consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9af: 2017-01-24T15:18:52.591Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0001): Fusing consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b1: 2017-01-24T15:18:52.593Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0d8b): Fusing consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b3: 2017-01-24T15:18:52.595Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0b15): Fusing consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b6: 2017-01-24T15:18:52.598Z: JOB_MESSAGE_DETAILED: (70e8f858d26e089f): Fusing consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9b9: 2017-01-24T15:18:52.601Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0629): Fusing consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9bc: 2017-01-24T15:18:52.604Z: JOB_MESSAGE_DETAILED: (70e8f858d26e03b3): Fusing consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9be: 2017-01-24T15:18:52.606Z: JOB_MESSAGE_DETAILED: (70e8f858d26e013d): Fusing consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10ee9c0: 2017-01-24T15:18:52.608Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0ec7): Fusing consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eea6c: 2017-01-24T15:18:52.780Z: JOB_MESSAGE_DEBUG: (70e8f858d26e0143): Workflow config is missing a default resource spec.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eea7d: 2017-01-24T15:18:52.797Z: JOB_MESSAGE_DETAILED: (70e8f858d26e0ecd): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eead2: 2017-01-24T15:18:52.882Z: JOB_MESSAGE_DEBUG: (a0290f97a19667a2): Adding workflow start and stop steps.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeb04: 2017-01-24T15:18:52.932Z: JOB_MESSAGE_DEBUG: (c297fa4537c49501): Assigning stage ids.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeb94: 2017-01-24T15:18:53.076Z: JOB_MESSAGE_DEBUG: (70e8f858d26e09e1): Executing wait step start2
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeba6: 2017-01-24T15:18:53.094Z: JOB_MESSAGE_DEBUG: (65a3ce102518fe29): Executing operation Some Numbers
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebb3: 2017-01-24T15:18:53.107Z: JOB_MESSAGE_DEBUG: (2eb55d8630d2cd6a): Value "Some Numbers.out" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebc1: 2017-01-24T15:18:53.121Z: JOB_MESSAGE_BASIC: S01: (2eb55d8630d2c79b): Executing operation assert:odd/Group/Create
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebc4: 2017-01-24T15:18:53.124Z: JOB_MESSAGE_BASIC: S03: (9a4f486fa95aca6f): Executing operation assert_that/Group/Create
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eebd1: 2017-01-24T15:18:53.137Z: JOB_MESSAGE_BASIC: S02: (2ab7216f4602c8a3): Executing operation assert:even/Group/Create
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eec8e: 2017-01-24T15:18:53.326Z: JOB_MESSAGE_DEBUG: (35722c8b1092de91): Starting worker pool setup.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eec90: 2017-01-24T15:18:53.328Z: JOB_MESSAGE_BASIC: (35722c8b1092d193): Starting 1 workers...
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeca5: 2017-01-24T15:18:53.349Z: JOB_MESSAGE_DEBUG: (70e8f858d26e04f5): Value "assert:odd/Group/Session" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeca7: 2017-01-24T15:18:53.351Z: JOB_MESSAGE_DEBUG: (5ca6d1bc2bc714ff): Value "assert:even/Group/Session" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eeca9: 2017-01-24T15:18:53.353Z: JOB_MESSAGE_DEBUG: (940ba07495593a6c): Value "assert_that/Group/Session" materialized.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d10eecb7: 2017-01-24T15:18:53.367Z: JOB_MESSAGE_BASIC: S04: (5ca6d1bc2bc71139): Executing operation ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d11092c8: 2017-01-24T15:20:41.416Z: JOB_MESSAGE_DETAILED: (45b4fae73847d61d): Workers have started successfully.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1114f4b: 2017-01-24T15:21:29.675Z: JOB_MESSAGE_ERROR: (51b0ae8abdf9362e): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d111504a: 2017-01-24T15:21:29.930Z: JOB_MESSAGE_ERROR: (51b0ae8abdf934a5): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d11150f7: 2017-01-24T15:21:30.103Z: JOB_MESSAGE_ERROR: (51b0ae8abdf9331c): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d111517f: 2017-01-24T15:21:30.239Z: JOB_MESSAGE_ERROR: (51b0ae8abdf936da): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115228: 2017-01-24T15:21:30.408Z: JOB_MESSAGE_ERROR: (51b0ae8abdf93551): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in _import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py", line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115254: 2017-01-24T15:21:30.452Z: JOB_MESSAGE_DEBUG: (6c3718a70c324aa): Executing failure step failure1
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115258: 2017-01-24T15:21:30.456Z: JOB_MESSAGE_ERROR: (6c3718a70c32f94): Workflow failed. Causes: (5ca6d1bc2bc7190c): S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write failed.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115299: 2017-01-24T15:21:30.521Z: JOB_MESSAGE_DETAILED: (66844a3ff238a206): Cleaning up.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115313: 2017-01-24T15:21:30.643Z: JOB_MESSAGE_DEBUG: (66844a3ff238a254): Starting worker pool teardown.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1115315: 2017-01-24T15:21:30.645Z: JOB_MESSAGE_BASIC: (66844a3ff238a2a2): Stopping worker pool...
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d112648c: 2017-01-24T15:22:40.652Z: JOB_MESSAGE_BASIC: (66844a3ff238ab17): Worker pool stopped.
root: INFO: 2017-01-24_07_18_48-4572845589957073573_00000159d1126893: 2017-01-24T15:22:41.683Z: JOB_MESSAGE_DEBUG: (66844a3ff238ac01): Tearing down pending resources...
root: INFO: Job 2017-01-24_07_18_48-4572845589957073573 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1024.260s

FAILED (errors=4)
Build step 'Execute shell' marked build as failure