You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by Dmytro Sen <ds...@hortonworks.com> on 2015/04/20 17:16:41 UTC
Review Request 33361: HDFS namenode_opt_newsize and
namenode_opt_maxnewsize not using number of datanodes in calculation
-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/33361/
-----------------------------------------------------------
Review request for Ambari, Myroslav Papirkovskyy and Srimanth Gunturi.
Bugs: AMBARI-10602
https://issues.apache.org/jira/browse/AMBARI-10602
Repository: ambari
Description
-------
Reviewing HDFS recommendation code for namenode_heapsize, namenode_opt_newsize and namenode_opt_maxnewsize, it looks like the calculations are not taking into account the number of datanodes and disks per datanode.
HDFS recommendations currently gives the following output
"configurations": {
"hadoop-env": {
"properties": {
"namenode_heapsize": "1024",
"namenode_opt_maxnewsize": "256",
"namenode_opt_newsize": "256"
},
"property_attributes": {
"namenode_heapsize": {
"maximum": "2887"
},
"dtnode_heapsize": {
"maximum": "2887"
}
}
},
Diffs
-----
ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py 860f6a2
ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py c0e5351
ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py 1f5549b
ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py 09a1f12
Diff: https://reviews.apache.org/r/33361/diff/
Testing
-------
----------------------------------------------------------------------
Ran 263 tests in 8.238s
OK
----------------------------------------------------------------------
Total run:728
Total errors:0
Total failures:0
OK
Thanks,
Dmytro Sen
Re: Review Request 33361: HDFS namenode_opt_newsize and
namenode_opt_maxnewsize not using number of datanodes in calculation
Posted by Myroslav Papirkovskyy <mp...@hortonworks.com>.
-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/33361/#review80927
-----------------------------------------------------------
Ship it!
Ship It!
- Myroslav Papirkovskyy
On Квітень 20, 2015, 6:16 після полудня, Dmytro Sen wrote:
>
> -----------------------------------------------------------
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/33361/
> -----------------------------------------------------------
>
> (Updated Квітень 20, 2015, 6:16 після полудня)
>
>
> Review request for Ambari, Myroslav Papirkovskyy and Srimanth Gunturi.
>
>
> Bugs: AMBARI-10602
> https://issues.apache.org/jira/browse/AMBARI-10602
>
>
> Repository: ambari
>
>
> Description
> -------
>
> Reviewing HDFS recommendation code for namenode_heapsize, namenode_opt_newsize and namenode_opt_maxnewsize, it looks like the calculations are not taking into account the number of datanodes and disks per datanode.
> HDFS recommendations currently gives the following output
> "configurations": {
> "hadoop-env": {
> "properties": {
> "namenode_heapsize": "1024",
> "namenode_opt_maxnewsize": "256",
> "namenode_opt_newsize": "256"
> },
> "property_attributes": {
> "namenode_heapsize": {
> "maximum": "2887"
> },
> "dtnode_heapsize": {
> "maximum": "2887"
> }
> }
> },
>
>
> Diffs
> -----
>
> ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py 860f6a2
> ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py c0e5351
> ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py 1f5549b
> ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py 09a1f12
>
> Diff: https://reviews.apache.org/r/33361/diff/
>
>
> Testing
> -------
>
> ----------------------------------------------------------------------
> Ran 263 tests in 8.238s
>
> OK
> ----------------------------------------------------------------------
> Total run:728
> Total errors:0
> Total failures:0
> OK
>
>
> Thanks,
>
> Dmytro Sen
>
>