You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Ahmet Altay (Jira)" <ji...@apache.org> on 2021/09/24 22:47:00 UTC
[jira] [Assigned] (BEAM-12958) All "Wordcount Dataflow" GHA checks
failing with PERMISSION_DENIED
[ https://issues.apache.org/jira/browse/BEAM-12958?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Ahmet Altay reassigned BEAM-12958:
----------------------------------
Assignee: Valentyn Tymofieiev
> All "Wordcount Dataflow" GHA checks failing with PERMISSION_DENIED
> ------------------------------------------------------------------
>
> Key: BEAM-12958
> URL: https://issues.apache.org/jira/browse/BEAM-12958
> Project: Beam
> Issue Type: Improvement
> Components: test-failures
> Reporter: Brian Hulette
> Assignee: Valentyn Tymofieiev
> Priority: P1
>
> "Wordcount Dataflow" checks for [Java|https://github.com/apache/beam/actions/runs/1268237569] and [Python|https://github.com/apache/beam/runs/3695202750?check_suite_focus=true] have been failing for some time.
> Error from Python:
> {code}
> 4s
> Run python -m apache_beam.examples.wordcount \
> INFO:apache_beam.internal.gcp.auth:Setting socket default timeout to 60 seconds.
> INFO:apache_beam.internal.gcp.auth:socket default timeout is 60.0 seconds.
> INFO:oauth2client.transport:Attempting refresh to obtain initial access_token
> INFO:oauth2client.client:Refreshing access_token
> INFO:apache_beam.runners.portability.stager:Copying Beam SDK "../../apache-beam-source/apache-beam-source.tar.gz" to staging location.
> WARNING:root:Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
> INFO:root:Default Python SDK image for environment is apache/beam_python3.6_sdk:2.34.0.dev
> INFO:root:Using provided Python SDK container image: gcr.io/cloud-dataflow/v1beta3/python36:beam-master-20210809
> INFO:root:Python SDK container image set to "gcr.io/cloud-dataflow/v1beta3/python36:beam-master-20210809" for Docker environment
> INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7fd41a43ad08> ====================
> INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7fd41a438510> ====================
> INFO:apache_beam.runners.dataflow.internal.apiclient:Defaulting to the temp_location as staging_location: gs://***/tmp/python_wordcount_dataflow/
> INFO:apache_beam.runners.dataflow.internal.apiclient:Starting GCS upload to gs://***/tmp/python_wordcount_dataflow/beamapp-runner-0924021925-749470.1632449965.749799/pickled_main_session...
> INFO:apache_beam.runners.dataflow.internal.apiclient:Completed GCS upload to gs://***/tmp/python_wordcount_dataflow/beamapp-runner-0924021925-749470.1632449965.749799/pickled_main_session in 0 seconds.
> INFO:apache_beam.runners.dataflow.internal.apiclient:Starting GCS upload to gs://***/tmp/python_wordcount_dataflow/beamapp-runner-0924021925-749470.1632449965.749799/dataflow_python_sdk.tar...
> INFO:apache_beam.runners.dataflow.internal.apiclient:Completed GCS upload to gs://***/tmp/python_wordcount_dataflow/beamapp-runner-0924021925-749470.1632449965.749799/dataflow_python_sdk.tar in 0 seconds.
> INFO:apache_beam.runners.dataflow.internal.apiclient:Starting GCS upload to gs://***/tmp/python_wordcount_dataflow/beamapp-runner-0924021925-749470.1632449965.749799/pipeline.pb...
> INFO:apache_beam.runners.dataflow.internal.apiclient:Completed GCS upload to gs://***/tmp/python_wordcount_dataflow/beamapp-runner-0924021925-749470.1632449965.749799/pipeline.pb in 0 seconds.
> Traceback (most recent call last):
> File "/opt/hostedtoolcache/Python/3.6.15/x64/lib/python3.6/runpy.py", line 193, in _run_module_as_main
> "__main__", mod_spec)
> File "/opt/hostedtoolcache/Python/3.6.15/x64/lib/python3.6/runpy.py", line 85, in _run_code
> exec(code, run_globals)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/examples/wordcount.py", line 94, in <module>
> run()
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/examples/wordcount.py", line 89, in run
> output | 'Write' >> WriteToText(known_args.output)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/pipeline.py", line 590, in __exit__
> self.result = self.run()
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/pipeline.py", line 543, in run
> self._options).run(False)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/pipeline.py", line 567, in run
> return self.runner.run_pipeline(self, self._options)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py", line 595, in run_pipeline
> self.dataflow_client.create_job(self.job), self)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/utils/retry.py", line 253, in wrapper
> return fun(*args, **kwargs)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py", line 695, in create_job
> return self.submit_job_description(job)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/utils/retry.py", line 253, in wrapper
> return fun(*args, **kwargs)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py", line 796, in submit_job_description
> response = self._client.projects_locations_jobs.Create(request)
> File "/home/runner/work/beam/beam/sdks/python/apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py", line 903, in Create
> config, request, global_params=global_params)
> File "/opt/hostedtoolcache/Python/3.6.15/x64/lib/python3.6/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
> return self.ProcessHttpResponse(method_config, http_response, request)
> File "/opt/hostedtoolcache/Python/3.6.15/x64/lib/python3.6/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
> self.__ProcessHttpResponse(method_config, http_response, request))
> File "/opt/hostedtoolcache/Python/3.6.15/x64/lib/python3.6/site-packages/apitools/base/py/base_api.py", line 604, in __ProcessHttpResponse
> http_response, method_config=method_config, request=request)
> apitools.base.py.exceptions.HttpForbiddenError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/***/locations/***/jobs?alt=json>: response: <***'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Fri, 24 Sep 2021 02:19:26 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'alt-svc': 'h3=":443"; ma=2592000,h3-29=":443"; ma=2592000,h3-T051=":443"; ma=2592000,h3-Q050=":443"; ma=2592000,h3-Q046=":443"; ma=2592000,h3-Q043=":443"; ma=2592000,quic=":443"; ma=2592000; v="46,43"', 'transfer-encoding': 'chunked', 'status': '403', 'content-length': '482', '-content-encoding': 'gzip'***>, content <***
> "error": ***
> "code": 403,
> "message": "(8264fa821ebdd451): Current user cannot act as service account 844138762903-compute@developer.gserviceaccount.com. Enforced by Org Policy constraint constraints/dataflow.enforceComputeDefaultServiceAccountCheck. https://cloud.google.com/iam/docs/service-accounts-actas Causes: (8264fa821ebddfa9): Current user cannot act as service account 844138762903-compute@developer.gserviceaccount.com.",
> "status": "PERMISSION_DENIED"
> ***
> ***
> >
> {code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)