You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Stefan Groschupf (JIRA)" <ji...@apache.org> on 2008/04/04 21:23:24 UTC

[jira] Updated: (HADOOP-1860) I'd like a log4j appender that can write to a Hadoop FileSystem

     [ https://issues.apache.org/jira/browse/HADOOP-1860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Stefan Groschupf updated HADOOP-1860:
-------------------------------------

    Attachment: hadoop-log4j.zip

I worked a little on that, though I never finished completely it, however it kind of works. Anyhow if someone ever stumble over  this issue I though it might be helpful to provide my code as starting point. The problem that is unsolved is the chicken egg problem, you need a running hdfs to write logs during startup. So you could use the appender and configure it to write into a log HDFS that already runs. 
However please be warned, writing logs into hdfs significant slows down hadoop!



> I'd like a log4j appender that can write to a Hadoop FileSystem
> ---------------------------------------------------------------
>
>                 Key: HADOOP-1860
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1860
>             Project: Hadoop Core
>          Issue Type: New Feature
>          Components: fs
>            Reporter: Owen O'Malley
>            Assignee: Owen O'Malley
>         Attachments: hadoop-log4j.zip
>
>
> It would be convenient to be able to write log files to HDFS and other file systems directly. For large clusters, it will produce too many files, but for small clusters it should be usable.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.