You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Jing Zhao (JIRA)" <ji...@apache.org> on 2014/05/17 00:36:14 UTC
[jira] [Created] (HDFS-6423) Diskspace quota usage is wrongly
updated when appending data from partial block
Jing Zhao created HDFS-6423:
-------------------------------
Summary: Diskspace quota usage is wrongly updated when appending data from partial block
Key: HDFS-6423
URL: https://issues.apache.org/jira/browse/HDFS-6423
Project: Hadoop HDFS
Issue Type: Bug
Reporter: Jing Zhao
Assignee: Jing Zhao
When appending new data to a file whose last block is a partial block, the diskspace quota usage is not correctly update. For example, suppose the block size is 1024 bytes, and a file has size 1536 bytes (1.5 blocks). If we then append another 1024 bytes to the file, the diskspace usage for this file will not be updated to (2560 * replication) as expected, but (2048 * replication).
The cause of the issue is that in FSNamesystem#commitOrCompleteLastBlock, we have
{code}
// Adjust disk space consumption if required
final long diff = fileINode.getPreferredBlockSize() - commitBlock.getNumBytes();
if (diff > 0) {
try {
String path = fileINode.getFullPathName();
dir.updateSpaceConsumed(path, 0, -diff*fileINode.getFileReplication());
} catch (IOException e) {
LOG.warn("Unexpected exception while updating disk space.", e);
}
}
{code}
This code assumes that the last block of the file has never been completed before, thus is always counted with the preferred block size in quota computation.
--
This message was sent by Atlassian JIRA
(v6.2#6252)