You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Myroslav Papirkovskyy (JIRA)" <ji...@apache.org> on 2015/04/29 18:12:06 UTC
[jira] [Created] (AMBARI-10837) HDFS Review: Multiple
recommendation API updates for HDFS configs
Myroslav Papirkovskyy created AMBARI-10837:
----------------------------------------------
Summary: HDFS Review: Multiple recommendation API updates for HDFS configs
Key: AMBARI-10837
URL: https://issues.apache.org/jira/browse/AMBARI-10837
Project: Ambari
Issue Type: Bug
Components: ambari-server
Affects Versions: 2.1.0
Reporter: Myroslav Papirkovskyy
Assignee: Myroslav Papirkovskyy
Priority: Critical
Fix For: 2.1.0
HDFS configs review was done and the configs spreadsheet has been updated with various changes where the following must be fixed.
* Below configs are to be marked as {{depends_on}} {{namenode_heapsize}}, and their value should be derived from it (basically ignore document value for below configs). Whenever {{namenode_heapsize}} changes in UI, below config values should also be updated
** namenode_opt_newsize (hadoop-env.sh) = {{namenode_heapsize/8}}
** namenode_opt_maxnewsize (hadoop-env.sh) = {{namenode_heapsize/8}}
* {{dfs.namenode.safemode.threshold-pct}}
** minimum = 0.990f
** maximum = 1.000f
** default = 0.999f
** increment-step = 0.001f
* {{dfs.datanode.failed.volumes.tolerated}} should be {{depends_on}} {{dfs.datanode.data.dir}}. So if a user adds additional folder in {{dfs.datanode.data.dir}}, then the *value and maximum* of {{dfs.datanode.failed.volumes.tolerated}} should change accordingly.
* {{namenode_heapsize}} calculations should take into account host memory limits. Namenode_heapsize should be the {{host-memory - os-reserverd-memory}}. Also, if there are any other master-components on the same host, then it should be halved (namenode_heapsize/2).
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)