You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@flume.apache.org by Juhani Connolly <ju...@gmail.com> on 2013/11/22 09:38:31 UTC
Review Request 15779: Flume-2245 HDFS Sink BucketWriter failing to close
after datanode issues
-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/15779/
-----------------------------------------------------------
Review request for Flume.
Repository: flume-git
Description
-------
https://issues.apache.org/jira/browse/FLUME-2245
Originally the flush() seemed superfluous however without it one of the unit tests breaks.
By moving on beyond regardless of the flush succeeding or not we allow the backing stream to actually get closed and reopened. While the real problem is with the HDFS stream not recovering this workaround seems necessary as otherwise appends will continue to fail until a restart.
Similarly HDFSDataStream and HDFSCompressedDataStream are closed regardless of the success of serialization/flushing. The exception should be propagated and cause a rollback so no data loss occurs.
Diffs
-----
flume-ng-sinks/flume-hdfs-sink/src/main/java/org/apache/flume/sink/hdfs/BucketWriter.java 200d457
flume-ng-sinks/flume-hdfs-sink/src/main/java/org/apache/flume/sink/hdfs/HDFSCompressedDataStream.java 5518547
flume-ng-sinks/flume-hdfs-sink/src/main/java/org/apache/flume/sink/hdfs/HDFSDataStream.java e20d1ee
Diff: https://reviews.apache.org/r/15779/diff/
Testing
-------
Existing unit tests pass.
I'm still trying to figure out a way to recreate the issue as it is hard to determine the exact cause
Thanks,
Juhani Connolly