You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2016/12/12 16:29:41 UTC
Build failed in Jenkins: beam_PostCommit_Python_Verify #864
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/864/>
------------------------------------------
[...truncated 2714 lines...]
},
{
"kind": "CollectionToSingleton",
"name": "s3",
"properties": {
"display_data": [],
"output_info": [
{
"encoding": {
"@type": "WindowedValueCoder$<string of 192 bytes>",
"component_encodings": [
{
"@type": "WindowedValueCoder$<string of 192 bytes>",
"component_encodings": [
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": [
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
},
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==",
"component_encodings": []
},
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": [
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
},
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
}
],
"is_pair_like": true
}
],
"is_wrapper": true
}
]
},
"output_name": "out",
"user_name": "ViewAsSingleton(side.None)/CreatePCollectionView.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s2"
},
"user_name": "ViewAsSingleton(side.None)/CreatePCollectionView"
}
},
{
"kind": "ParallelDo",
"name": "s4",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "<lambda>"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {
"s3": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s3"
}
},
"output_info": [
{
"encoding": {
"@type": "WindowedValueCoder$<string of 408 bytes>",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==",
"component_encodings": []
},
{
"@type": "SingletonCoder$<string of 256 bytes>",
"component_encodings": []
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "compute.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s1"
},
"serialized_fn": "<string of 1116 bytes>",
"user_name": "compute"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
id: u'2016-12-12_07_50_24-6463664622346895841'
projectId: u'apache-beam-testing'
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2016-12-12_07_50_24-6463664622346895841]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2016-12-12_07_50_24-6463664622346895841
root: INFO: Job 2016-12-12_07_50_24-6463664622346895841 is in state JOB_STATE_RUNNING
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5ac0: 2016-12-12T15:50:26.240Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28e44): Checking required Cloud APIs are enabled.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c3d: 2016-12-12T15:50:26.621Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28d09): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c3f: 2016-12-12T15:50:26.623Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28cc3): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c42: 2016-12-12T15:50:26.626Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28c7d): Annotating graph with Autotuner information.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c4b: 2016-12-12T15:50:26.635Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28bab): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c5b: 2016-12-12T15:50:26.651Z: JOB_MESSAGE_DEBUG: (a4af248cc1e289c1): Workflow config is missing a default resource spec.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c5d: 2016-12-12T15:50:26.653Z: JOB_MESSAGE_DETAILED: (a4af248cc1e2897b): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c98: 2016-12-12T15:50:26.712Z: JOB_MESSAGE_DEBUG: (7e90f8ebe8480cde): Adding workflow start and stop steps.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5cca: 2016-12-12T15:50:26.762Z: JOB_MESSAGE_DEBUG: (d8a1c8535c37ea18): Assigning stage ids.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d36: 2016-12-12T15:50:26.870Z: JOB_MESSAGE_DEBUG: (877e4ceff4313187): Executing wait step start2
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d3f: 2016-12-12T15:50:26.879Z: JOB_MESSAGE_DEBUG: (31cd13e4fae54049): Executing operation side
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d4b: 2016-12-12T15:50:26.891Z: JOB_MESSAGE_DEBUG: (263b1beb2e4eb699): Value "side.out" materialized.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d5f: 2016-12-12T15:50:26.911Z: JOB_MESSAGE_BASIC: S01: (81ad196c466d6570): Executing operation ViewAsSingleton(side.None)/CreatePCollectionView
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d65: 2016-12-12T15:50:26.917Z: JOB_MESSAGE_DEBUG: (d8a1c8535c37e4ba): Executing operation start
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d6c: 2016-12-12T15:50:26.924Z: JOB_MESSAGE_DEBUG: (7e90f8ebe8480d21): Value "ViewAsSingleton(side.None)/CreatePCollectionView.out" materialized.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d70: 2016-12-12T15:50:26.928Z: JOB_MESSAGE_DEBUG: (667056a065d57d76): Value "start.out" materialized.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d7c: 2016-12-12T15:50:26.940Z: JOB_MESSAGE_BASIC: S02: (263b1beb2e4eb07b): Executing operation compute
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5e47: 2016-12-12T15:50:27.143Z: JOB_MESSAGE_DEBUG: (98ee7be1df6055ce): Starting worker pool setup.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5e49: 2016-12-12T15:50:27.145Z: JOB_MESSAGE_BASIC: (98ee7be1df605c54): Starting 1 workers...
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bc3dc7: 2016-12-12T15:52:29.895Z: JOB_MESSAGE_DETAILED: (6607e90d3f03422d): Workers have started successfully.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd1134: 2016-12-12T15:53:24.020Z: JOB_MESSAGE_ERROR: (cc11e033844b1905): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd11a4: 2016-12-12T15:53:24.132Z: JOB_MESSAGE_ERROR: (cc11e033844b1835): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd121c: 2016-12-12T15:53:24.252Z: JOB_MESSAGE_ERROR: (cc11e033844b1765): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd127f: 2016-12-12T15:53:24.351Z: JOB_MESSAGE_ERROR: (cc11e033844b1695): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd12a4: 2016-12-12T15:53:24.388Z: JOB_MESSAGE_DEBUG: (263b1beb2e4ebe21): Executing failure step failure1
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd12a6: 2016-12-12T15:53:24.390Z: JOB_MESSAGE_ERROR: (263b1beb2e4ebc17): Workflow failed. Causes: (263b1beb2e4eb33a): S02:compute failed.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd12da: 2016-12-12T15:53:24.442Z: JOB_MESSAGE_DETAILED: (d8bd0f04c4d5c1ec): Cleaning up.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd1362: 2016-12-12T15:53:24.578Z: JOB_MESSAGE_DEBUG: (d8bd0f04c4d5c39b): Starting worker pool teardown.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd1364: 2016-12-12T15:53:24.580Z: JOB_MESSAGE_BASIC: (d8bd0f04c4d5cf65): Stopping worker pool...
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3be4bea: 2016-12-12T15:54:44.586Z: JOB_MESSAGE_BASIC: (d8bd0f04c4d5c114): Worker pool stopped.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3be4bfc: 2016-12-12T15:54:44.604Z: JOB_MESSAGE_DEBUG: (d8bd0f04c4d5ce8d): Tearing down pending resources...
root: INFO: Job 2016-12-12_07_50_24-6463664622346895841 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 21 tests in 5235.902s
FAILED (errors=1)
Build step 'Execute shell' marked build as failure
Jenkins build is back to normal : beam_PostCommit_Python_Verify
#866
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/866/changes>
Build failed in Jenkins: beam_PostCommit_Python_Verify #865
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/865/>
------------------------------------------
[...truncated 2748 lines...]
"component_encodings": [
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
},
{
"@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
}
],
"is_pair_like": true
}
],
"is_wrapper": true
}
]
},
"output_name": "out",
"user_name": "ViewAsSingleton(side.None)/CreatePCollectionView.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s2"
},
"user_name": "ViewAsSingleton(side.None)/CreatePCollectionView"
}
},
{
"kind": "ParallelDo",
"name": "s4",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "<lambda>"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {
"s3": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s3"
}
},
"output_info": [
{
"encoding": {
"@type": "WindowedValueCoder$<string of 408 bytes>",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==",
"component_encodings": []
},
{
"@type": "SingletonCoder$<string of 256 bytes>",
"component_encodings": []
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "compute.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s1"
},
"serialized_fn": "<string of 1116 bytes>",
"user_name": "compute"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
id: u'2016-12-12_14_37_00-14980560058396649741'
projectId: u'apache-beam-testing'
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2016-12-12_14_37_00-14980560058396649741]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2016-12-12_14_37_00-14980560058396649741
root: INFO: Job 2016-12-12_14_37_00-14980560058396649741 is in state JOB_STATE_RUNNING
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e97f6: 2016-12-12T22:37:01.302Z: JOB_MESSAGE_DETAILED: (d532e8081463d3f0): Checking required Cloud APIs are enabled.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9978: 2016-12-12T22:37:01.688Z: JOB_MESSAGE_DETAILED: (d532e8081463d065): Expanding GroupByKey operations into optimizable parts.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e997b: 2016-12-12T22:37:01.691Z: JOB_MESSAGE_DETAILED: (d532e8081463d87f): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e997d: 2016-12-12T22:37:01.693Z: JOB_MESSAGE_DETAILED: (d532e8081463d099): Annotating graph with Autotuner information.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9985: 2016-12-12T22:37:01.701Z: JOB_MESSAGE_DETAILED: (d532e8081463d8e7): Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9994: 2016-12-12T22:37:01.716Z: JOB_MESSAGE_DEBUG: (d532e8081463d19d): Workflow config is missing a default resource spec.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9996: 2016-12-12T22:37:01.718Z: JOB_MESSAGE_DETAILED: (d532e8081463d9b7): Adding StepResource setup and teardown to workflow graph.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e99a3: 2016-12-12T22:37:01.731Z: JOB_MESSAGE_DEBUG: (1bb8e82eeb952c4a): Adding workflow start and stop steps.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e99d0: 2016-12-12T22:37:01.776Z: JOB_MESSAGE_DEBUG: (e2e0641eabcd52bd): Assigning stage ids.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9a66: 2016-12-12T22:37:01.926Z: JOB_MESSAGE_DEBUG: (dcd6e9da78048dfd): Executing wait step start2
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9a72: 2016-12-12T22:37:01.938Z: JOB_MESSAGE_DEBUG: (dcd6e9da78048490): Executing operation start
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9a99: 2016-12-12T22:37:01.977Z: JOB_MESSAGE_DEBUG: (639f84ea478d1e9a): Value "start.out" materialized.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9a9d: 2016-12-12T22:37:01.981Z: JOB_MESSAGE_DEBUG: (b08cd86600a0c86b): Executing operation side
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9aa9: 2016-12-12T22:37:01.993Z: JOB_MESSAGE_DEBUG: (b08cd86600a0c411): Value "side.out" materialized.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9ab3: 2016-12-12T22:37:02.003Z: JOB_MESSAGE_BASIC: S01: (3dd7dcc5e5222773): Executing operation ViewAsSingleton(side.None)/CreatePCollectionView
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9ac0: 2016-12-12T22:37:02.016Z: JOB_MESSAGE_DEBUG: (6ed70e308ca37aca): Value "ViewAsSingleton(side.None)/CreatePCollectionView.out" materialized.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9ac9: 2016-12-12T22:37:02.025Z: JOB_MESSAGE_BASIC: S02: (3fe6eceb837205a0): Executing operation compute
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9b95: 2016-12-12T22:37:02.229Z: JOB_MESSAGE_DEBUG: (4fc3e19b5cd6b61c): Starting worker pool setup.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f52e9b97: 2016-12-12T22:37:02.231Z: JOB_MESSAGE_BASIC: (4fc3e19b5cd6b536): Starting 1 workers...
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5307e3f: 2016-12-12T22:39:05.791Z: JOB_MESSAGE_DETAILED: (33dc20d64b2fb257): Workers have started successfully.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314c7f: 2016-12-12T22:39:58.591Z: JOB_MESSAGE_ERROR: (363bf96903e34635): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314cdd: 2016-12-12T22:39:58.685Z: JOB_MESSAGE_ERROR: (363bf96903e34e05): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314d38: 2016-12-12T22:39:58.776Z: JOB_MESSAGE_ERROR: (363bf96903e345d5): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314d9c: 2016-12-12T22:39:58.876Z: JOB_MESSAGE_ERROR: (363bf96903e34da5): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314dfa: 2016-12-12T22:39:58.970Z: JOB_MESSAGE_ERROR: (363bf96903e34575): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314e57: 2016-12-12T22:39:59.063Z: JOB_MESSAGE_ERROR: (363bf96903e34d45): Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008)
op.start()
File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
def start(self):
File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
self.dofn_runner = common.DoFnRunner(
File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469)
args, kwargs, [side_input[global_window]
File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__
_FilteringIterable(self._iterable, target_window), self._view_options)
File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable
'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314e7b: 2016-12-12T22:39:59.099Z: JOB_MESSAGE_DEBUG: (d532e8081463d5f8): Executing failure step failure1
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314e7d: 2016-12-12T22:39:59.101Z: JOB_MESSAGE_ERROR: (d532e8081463de12): Workflow failed. Causes: (3fe6eceb83720023): S02:compute failed.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314eaf: 2016-12-12T22:39:59.151Z: JOB_MESSAGE_DETAILED: (d7e9bc71ab46918): Cleaning up.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314f39: 2016-12-12T22:39:59.289Z: JOB_MESSAGE_DEBUG: (d7e9bc71ab46967): Starting worker pool teardown.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5314f3b: 2016-12-12T22:39:59.291Z: JOB_MESSAGE_BASIC: (d7e9bc71ab46ef1): Stopping worker pool...
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5329b4a: 2016-12-12T22:41:24.298Z: JOB_MESSAGE_BASIC: (d7e9bc71ab46f40): Worker pool stopped.
root: INFO: 2016-12-12_14_37_00-14980560058396649741_00000158f5329b5a: 2016-12-12T22:41:24.314Z: JOB_MESSAGE_DEBUG: (d7e9bc71ab46519): Tearing down pending resources...
root: INFO: Job 2016-12-12_14_37_00-14980560058396649741 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 21 tests in 5341.580s
FAILED (errors=1)
Build step 'Execute shell' marked build as failure