You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by Apache Wiki <wi...@apache.org> on 2010/08/02 20:23:51 UTC

[Hadoop Wiki] Update of "Hbase/FAQ" by MaximVeksler

Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "Hbase/FAQ" page has been changed by MaximVeksler.
http://wiki.apache.org/hadoop/Hbase/FAQ?action=diff&rev1=66&rev2=67

--------------------------------------------------

  }}}
  '''''hbase''' is the user under which HBase is running''. To test the configuration reboot and run '''ulimit -n'''
  
- You may also need to edit /etc/sysctl.conf, relevant configuration '''fs.file-max''' See http://thedaneshproject.com/posts/how-to-increase-total-file-descriptors-count-on-linux/
+ You may also need to edit /etc/sysctl.conf, relevant configuration '''fs.file-max''' See http://serverfault.com/questions/165316/how-to-configure-linux-file-descriptor-limit-with-fs-file-max-and-ulimit/
  
  The math runs roughly as follows: Per column family, there is at least one mapfile and possibly up to 5 or 6 if a region is under load (lets say 3 per column family on average).  Multiply by the number of regions per region server.  So, for example, say you have a schema of 3 column familes per region and that you have 100 regions per regionserver, the JVM will open 3 * 3 * 100 mapfiles -- 900 file descriptors not counting open jar files, conf files, etc (Run 'lsof -p REGIONSERVER_PID' to see for sure).