You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Mark Liu (JIRA)" <ji...@apache.org> on 2018/11/30 01:02:00 UTC

[jira] [Created] (BEAM-6154) Gcsio batch delete broken in Python 3

Mark Liu created BEAM-6154:
------------------------------

             Summary: Gcsio batch delete broken in Python 3
                 Key: BEAM-6154
                 URL: https://issues.apache.org/jira/browse/BEAM-6154
             Project: Beam
          Issue Type: Bug
          Components: sdk-py-core
            Reporter: Mark Liu
            Assignee: Ahmet Altay


I'm running Python SDK agianst GCP in Python 3.5 and got following gcsio error while deleting files:

{code}
  File "/usr/local/lib/python3.5/site-packages/apache_beam/io/iobase.py", line 1077, in <genexpr>
    window.TimestampedValue(v, timestamp.MAX_TIMESTAMP) for v in outputs)
  File "/usr/local/lib/python3.5/site-packages/apache_beam/io/filebasedsink.py", line 315, in finalize_write
    num_threads)
  File "/usr/local/lib/python3.5/site-packages/apache_beam/internal/util.py", line 145, in run_using_threadpool
    return pool.map(fn_to_execute, inputs)
  File "/usr/local/lib/python3.5/multiprocessing/pool.py", line 266, in map
    return self._map_async(func, iterable, mapstar, chunksize).get()
  File "/usr/local/lib/python3.5/multiprocessing/pool.py", line 644, in get
    raise self._value
  File "/usr/local/lib/python3.5/multiprocessing/pool.py", line 119, in worker
    result = (True, func(*args, **kwds))
  File "/usr/local/lib/python3.5/multiprocessing/pool.py", line 44, in mapstar
    return list(map(*args))
  File "/usr/local/lib/python3.5/site-packages/apache_beam/io/filebasedsink.py", line 299, in _rename_batch
    FileSystems.rename(source_files, destination_files)
  File "/usr/local/lib/python3.5/site-packages/apache_beam/io/filesystems.py", line 252, in rename
    return filesystem.rename(source_file_names, destination_file_names)
  File "/usr/local/lib/python3.5/site-packages/apache_beam/io/gcp/gcsfilesystem.py", line 229, in rename
    copy_statuses = gcsio.GcsIO().copy_batch(batch)
  File "/usr/local/lib/python3.5/site-packages/apache_beam/io/gcp/gcsio.py", line 322, in copy_batch
    api_calls = batch_request.Execute(self.client._http)  # pylint: disable=protected-access
  File "/usr/local/lib/python3.5/site-packages/apitools/base/py/batch.py", line 222, in Execute
    batch_http_request.Execute(http)
  File "/usr/local/lib/python3.5/site-packages/apitools/base/py/batch.py", line 480, in Execute
    self._Execute(http)
  File "/usr/local/lib/python3.5/site-packages/apitools/base/py/batch.py", line 450, in _Execute
    mime_response = parser.parsestr(header + response.content)
TypeError: Can't convert 'bytes' object to str implicitly
{code} 

After looking into related code in apitools library, I found response.content that's returned via http request to gcs is bytes and apitools didn't handle this scenario. This can be a blocker to any pipeline depending on gcsio and apparently blocks all Dataflow job in Python 3.

This could be another case that moving off apitools dependency in [BEAM-4850|https://issues.apache.org/jira/browse/BEAM-4850].



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)