You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by "Kamil Bregula (Jira)" <ji...@apache.org> on 2019/09/08 18:15:00 UTC
[jira] [Commented] (AIRFLOW-5418) bucket names with a slash in them
cause an error with the GCS Download Operator
[ https://issues.apache.org/jira/browse/AIRFLOW-5418?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16925226#comment-16925226 ]
Kamil Bregula commented on AIRFLOW-5418:
----------------------------------------
Slash in a bucket name is restricted?
{quote}Bucket names must contain only lowercase letters, numbers, dashes ({{-}}), underscores ({{_}}), and dots ({{.}}). Names containing dots require [verification|https://cloud.google.com/storage/docs/domain-name-verification].
{quote}
[https://cloud.google.com/storage/docs/naming]
> bucket names with a slash in them cause an error with the GCS Download Operator
> -------------------------------------------------------------------------------
>
> Key: AIRFLOW-5418
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5418
> Project: Apache Airflow
> Issue Type: Bug
> Components: operators
> Affects Versions: 1.8.0, 1.8.1, 1.8.2, 1.9.0, 1.10.0, 1.10.1, 1.10.2, 1.10.3, 1.10.4, 1.10.5
> Reporter: Leah
> Priority: Minor
>
> If you try to use the GCS Downlaod Operator on a bucket that has a slash in the name, for example, my-bucket/dags, the operation results in a "Bad Request" error. I replicated this error consistently in Cloud Composer (GCP hosted managed Airflow) and was able to get it to work on a bucket name without a slash in it. I was not able to use a backslash to escape the slash when passing in the bucket name either.
>
>
> {{[2019-09-05 23:51:53,951] \{models.py:1796} ERROR - <HttpError 400 when requesting https://www.googleapis.com/storage/v1/b/my-bucket/dags/o/gcs_download.py?alt=media returned "Bad Request">}}
--
This message was sent by Atlassian Jira
(v8.3.2#803003)