You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2017/05/27 01:08:24 UTC
Build failed in Jenkins: beam_PostCommit_Python_Verify #2336
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2336/display/redirect?page=changes>
Changes:
[chamikara] Fix Python Dataflow execution errors due to #3223
------------------------------------------
[...truncated 677.55 KB...]
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/funcsigs-1.0.2.tar.gz",
"name": "funcsigs-1.0.2.tar.gz"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/PyHamcrest-1.9.0.tar.gz",
"name": "PyHamcrest-1.9.0.tar.gz"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/pbr-1.10.0.tar.gz",
"name": "pbr-1.10.0.tar.gz"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/setuptools-35.0.0.zip",
"name": "setuptools-35.0.0.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/setuptools-34.3.1.zip",
"name": "setuptools-34.3.1.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/setuptools-34.3.3.zip",
"name": "setuptools-34.3.3.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/setuptools-34.3.2.zip",
"name": "setuptools-34.3.2.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/setuptools-34.1.1.zip",
"name": "setuptools-34.1.1.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/pyparsing-2.1.10.zip",
"name": "pyparsing-2.1.10.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/setuptools-34.0.2.zip",
"name": "setuptools-34.0.2.zip"
},
{
"location": "storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0527004503-490606.1495845903.490825/dataflow_python_sdk.tar",
"name": "dataflow_python_sdk.tar"
}
],
"taskrunnerSettings": {
"parallelWorkerSettings": {
"baseUrl": "https://dataflow.googleapis.com",
"servicePath": "https://dataflow.googleapis.com"
}
},
"workerHarnessContainerImage": "dataflow.gcr.io/v1beta3/python:beam-2.1.0-20170518"
}
]
},
"name": "beamapp-jenkins-0527004503-490606",
"steps": [
{
"kind": "ParallelRead",
"name": "s1",
"properties": {
"custom_source_step_input": {
"metadata": {
"estimated_size_bytes": {
"@type": "http://schema.org/Integer",
"value": 100
}
},
"spec": {
"@type": "CustomSourcesType",
"serialized_source": "<string of 292 bytes>"
}
},
"display_data": [
{
"key": "source",
"label": "Read Source",
"namespace": "apache_beam.io.iobase.Read",
"shortValue": "CountingSource",
"type": "STRING",
"value": "apache_beam.examples.snippets.snippets.CountingSource"
}
],
"format": "custom_source",
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "PickleCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": [
{
"@type": "PickleCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
},
{
"@type": "PickleCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
"component_encodings": []
}
],
"is_pair_like": true
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "Read.out"
}
],
"user_name": "Read"
}
},
{
"kind": "ParallelDo",
"name": "s2",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CounterDoFn",
"type": "STRING",
"value": "apache_beam.transforms.ptransform_test.CounterDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "ParDo(CounterDoFn).out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s1"
},
"serialized_fn": "<string of 1152 bytes>",
"user_name": "ParDo(CounterDoFn)"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Sat, 27 May 2017 00:45:21 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(55d84ae3691726e7): The workflow could not be created. Causes: (e8dc10538ead1ee1): Too many running jobs. Project apache-beam-testing is running 26 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Sat, 27 May 2017 00:45:23 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(36ce906bca41c0ee): The workflow could not be created. Causes: (1f7f41a6405bbb8e): Too many running jobs. Project apache-beam-testing is running 26 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Sat, 27 May 2017 00:45:27 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(3637d2126f244f89): The workflow could not be created. Causes: (bcb6ddc378cc2049): Too many running jobs. Project apache-beam-testing is running 26 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Sat, 27 May 2017 00:45:37 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(6234c6ebbf479e5b): The workflow could not be created. Causes: (a778c6a28225564a): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: ERROR: HTTP status 429 trying to create job at dataflow service endpoint https://dataflow.googleapis.com
root: CRITICAL: details of server error: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '440', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Sat, 27 May 2017 00:45:57 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(9927637e08f0d01e): The workflow could not be created. Causes: (658a9e5fd754639): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedback@google.com to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 15 tests in 1400.072s
FAILED (errors=2)
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_45_36-10498212390368848663?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_51_37-17132228235523379593?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_57_48-17027123095381803203?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_18_02_56-10752183607397322683?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_45_35-6398202493305249974?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_51_41-2913020719255933115?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_57_23-823607875585256437?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_46_08-16322035139584704846?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_53_10-12803941393774394226?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_59_06-5336498007307815832?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_46_12-16259545247750750650?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_52_37-18110321671650162858?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-26_17_58_33-4360611796317224837?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user klk@google.com
Jenkins build is back to normal : beam_PostCommit_Python_Verify
#2337
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2337/display/redirect>