You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hadoop.apache.org by Deepak Nagaraj <n....@gmail.com> on 2016/10/03 19:58:03 UTC
du says 3X more used than dfsadmin
Hi,
I'm running a single-node test HDFS cluster (RF=1, in case that matters).
Here's my problem:
* hdfs -dfsadmin shows 10G for "dfsUsed".
* But "du -sh" on the data directory (/var/hadoop_data/hdfs/datanode) shows 30G.
I also noticed that the following file has around ~30G for dfsUsed:
$ cat dfsUsed
32516294656 1474677109034
$ pwd
/var/hadoop_data/hdfs/datanode/current/BP-<XYZ>/current
So I'm wondering if dfsadmin is not picking it up correctly?
The end result is that a script that's supposed to archive old data is
not doing anything, because it thinks there's more space, whereas in
reality there isn't any.
Can someone please help me here?
Thanks in advance,
-deepak
---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@hadoop.apache.org
For additional commands, e-mail: user-help@hadoop.apache.org