You are viewing a plain text version of this content. The canonical link for it is here.
Posted to jira@arrow.apache.org by "Nikolay Markov (Jira)" <ji...@apache.org> on 2021/11/12 14:57:00 UTC
[jira] [Commented] (ARROW-12338) [Python] Permission denied while
accessing HDFS data
[ https://issues.apache.org/jira/browse/ARROW-12338?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17442787#comment-17442787 ]
Nikolay Markov commented on ARROW-12338:
----------------------------------------
Having the same issue on pyarrow==5.0.0. Kerberos ticket and username are correct, everything works from command line. Trying to dig deeper, but still no luck.
> [Python] Permission denied while accessing HDFS data
> ----------------------------------------------------
>
> Key: ARROW-12338
> URL: https://issues.apache.org/jira/browse/ARROW-12338
> Project: Apache Arrow
> Issue Type: Bug
> Components: Python
> Affects Versions: 3.0.0
> Reporter: Suhas N M
> Priority: Blocker
> Labels: filesystem, hdfs
>
> Hi, I have been trying to connect to HDFS cluster using pyarrow version 3.0.0, connection goes through, but I am unable to perform any operation involving HDFS cluster. Here is the error thrown:
> Traceback (most recent call last):
> File "pyarrow_test.py", line 8, in <module>
> hdfs.create_dir('test3')
> File "pyarrow/_fs.pyx", line 450, in pyarrow._fs.FileSystem.create_dir
> File "pyarrow/error.pxi", line 99, in pyarrow.lib.check_status
> OSError: HDFS create directory failed, errno: 13 (Permission denied)
> PS: I have checked access permissions and they are correct. I am able to access the files and create directories with the 'hdfs' command.
> Hadoop cluster is Kerberos enabled, I have used the following line to create connection:
> hdfs = fs.HadoopFileSystem('<hostname>', 8020, user='<username>', kerb_ticket='/tmp/krb5cc_500')
--
This message was sent by Atlassian Jira
(v8.20.1#820001)