You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ranger.apache.org by "Ramesh Mani (Jira)" <ji...@apache.org> on 2021/12/08 07:50:00 UTC

[jira] [Updated] (RANGER-3271) Ranger Knox Plugin Unable to Write Knox Audits to HDFS

     [ https://issues.apache.org/jira/browse/RANGER-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Ramesh Mani updated RANGER-3271:
--------------------------------
    Fix Version/s: 2.1.0

> Ranger Knox Plugin Unable to Write  Knox Audits to HDFS
> -------------------------------------------------------
>
>                 Key: RANGER-3271
>                 URL: https://issues.apache.org/jira/browse/RANGER-3271
>             Project: Ranger
>          Issue Type: Bug
>          Components: Ranger
>    Affects Versions: 2.1.0
>         Environment: HADOOP : 3.3.0
> KNOX : 1.4.0
> RANGER : 2.1.0
>            Reporter: Venkat A
>            Priority: Blocker
>             Fix For: 2.1.0
>
>
> I see following error when Knox audits being written to HDFS after Ranger-Knox plugin enabled.
>  
>  3322021-05-01 15:33:34,435 INFO destination.HDFSAuditDestination: Returning HDFS Filesystem Config: Configuration: core-default.xml, core-site.xml 3322021-05-01 15:33:34,435 INFO destination.HDFSAuditDestination: Returning HDFS Filesystem Config: Configuration: core-default.xml, core-site.xml 3182021-05-01 15:33:34,439 ERROR provider.BaseAuditHandler: Error writing to log file. 349{color:#FF0000}*org.apache.hadoop.fs.UnsupportedFileSystemException: No FileSystem for scheme "hdfs"*{color} at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:3332) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3352) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477) at org.apache.ranger.audit.destination.HDFSAuditDestination.getLogFileStream(HDFSAuditDestination.java:277) at org.apache.ranger.audit.destination.HDFSAuditDestination.access$000(HDFSAuditDestination.java:44) at org.apache.ranger.audit.destination.HDFSAuditDestination$1.run(HDFSAuditDestination.java:157) at org.apache.ranger.audit.destination.HDFSAuditDestination$1.run(HDFSAuditDestination.java:154) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) at org.apache.ranger.audit.provider.MiscUtil.executePrivilegedAction(MiscUtil.java:529) at org.apache.ranger.audit.destination.HDFSAuditDestination.logJSON(HDFSAuditDestination.java:154) at org.apache.ranger.audit.queue.AuditFileSpool.sendEvent(AuditFileSpool.java:879) at org.apache.ranger.audit.queue.AuditFileSpool.runLogAudit(AuditFileSpool.java:827) at org.apache.ranger.audit.queue.AuditFileSpool.run(AuditFileSpool.java:757) at java.lang.Thread.run(Thread.java:748)2021-05-01 15:33:34,439 INFO destination.HDFSAuditDestination: Flushing HDFS audit. Event Size:1 1802021-05-01 15:33:34,439 ERROR queue.AuditFileSpool: Error sending logs to consumer. provider=knox.async.multi_dest.batch, consumer=knox.async.multi_dest.batch.hdfs 7092021-05-01 15:33:34,440 INFO queue.AuditFileSpool: Destination is down. sleeping for 30000 milli seconds. indexQueue=0, queueName=knox.async.multi_dest.batch, consumer=knox.async.multi_dest.batch.hdfs
>  
>  
> Tried lot of options to avoid above error. Not sure if it is a bug or some sort of Compatibility issue.
> Environment : 
> HADOOP : 3.3.0
> KNOX : 1.4.0
> RANGER : 2.1.0
> NOTE: 
> Knox able to write audits if i give local path to store audits instead of HDFS File System.
> Appreciate your help on this.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)