You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "Elliott Clark (JIRA)" <ji...@apache.org> on 2013/08/26 20:43:52 UTC

[jira] [Commented] (HBASE-9338) Test Big Linked List fails on Hadoop 2.1.0

    [ https://issues.apache.org/jira/browse/HBASE-9338?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13750383#comment-13750383 ] 

Elliott Clark commented on HBASE-9338:
--------------------------------------

Configs on the cluster:

core-site.xml
{code}
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>

  <property>
    <name>fs.default.name</name>
    <value>hdfs://${MASTER_HOSTNAME}:8020/</value>
  </property>

  <property>
    <name>hadoop.data.dir0</name>
    <value>/data0/</value>
  </property>

  <property>
    <name>hadoop.data.dir1</name>
    <value>/data1/</value>
  </property>

  <property>
    <name>hadoop.data.dir2</name>
    <value>/data2/</value>
  </property>

  <property>
    <name>hadoop.data.dir3</name>
    <value>/data3/</value>
  </property>

  <property>
    <name>hadoop.data.dir4</name>
    <value>/data4/</value>
  </property>

  <property>
    <name>hadoop.data.dir5</name>
    <value>/data5/</value>
  </property>

  <property>
    <name>mapred.temp.dir</name>
    <value>${hadoop.data.dir1}/mapred/temp</value>
    <description>A shared directory for temporary files.
    </description>
  </property>

  <property>
    <name>ipc.client.connect.timeout</name>
    <value>1000</value>
  </property>

  <property>
    <name>ipc.client.connect.max.retries.on.timeouts</name>
    <value>2</value>
  </property>

  <property>
    <name>dfs.socket.timeout</name>
    <value>5000</value>
  </property>

  <property>
    <name>dfs.socket.write.timeout</name>
    <value>5000</value>
  </property>

  <property>
    <name>ipc.ping.interval</name>
    <value>20000</value>
  </property>

  <property>
    <name>io.file.buffer.size</name>
    <value>65536</value>
  </property>

  <property>
    <name>ipc.client.connect.max.retries</name>
    <value>10</value>
  </property>

  <property>
    <name>ipc.client.tcpnodelay</name>
    <value>true</value>
  </property>

  <property>
    <name>ipc.server.tcpnodelay</name>
    <value>true</value>
  </property>

  <property>
    <name>dfs.domain.socket.path</name>
    <value>/var/lib/hadoop/dn_socket._PORT</value>
  </property>
  
  <property>
    <name>dfs.client.read.shortcircuit</name>
    <value>true</value>
  </property>

  <property>
    <name>dfs.client.file-block-storage-locations.timeout</name>
    <value>3000</value>
  </property>

  <property>
     <name>dfs.client.read.shortcircuit.skip.checksum</name>
     <value>true</value>
   </property>

</configuration>
{code}

hdfs-site.xml
{code}
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<configuration>

  <property>
    <name>dfs.datanode.handler.count</name>
    <!-- default 10 -->
    <value>32</value>
    <description>The number of server threads for the
    datanode.</description>
  </property>

  <property>
    <name>dfs.namenode.handler.count</name>
    <!-- default 10 -->
    <value>32</value>
    <description>The number of server threads for the
    namenode.</description>
  </property>

  <property>
    <name>dfs.block.size</name>
    <value>134217728</value>
    <description>The default block size for new files.</description>
  </property>

  <property>
    <name>dfs.datanode.max.xcievers</name>
    <value>4098</value>
  </property>

  <property>
    <name>dfs.namenode.replication.interval</name>
    <value>15</value>
  </property>

  <property>
    <name>dfs.balance.bandwidthPerSec</name>
    <value>10485760</value>
  </property>

  <property>
    <name>fs.checkpoint.dir</name>
    <value>${hadoop.data.dir1}/dfs/namesecondary</value>
  </property>

  <property>
    <name>dfs.name.dir</name>
    <value>${hadoop.data.dir0}/dfs/name</value>
  </property>

  <property>
    <name>dfs.data.dir</name>
    <value>${hadoop.data.dir0}/dfs/data,${hadoop.data.dir1}/dfs/data,${hadoop.data.dir2}/dfs/data,${hadoop.data.dir3}/dfs/data,${hadoop.data.dir4}/dfs/data,${hadoop.data.dir5}/dfs/data</value>
  </property>

  <property>
    <name>dfs.datanode.socket.write.timeout</name>
    <value>10000</value>
  </property>

  <property>
    <name>ipc.client.connect.timeout</name>
    <value>1000</value>
  </property>

  <property>
    <name>ipc.client.connect.max.retries.on.timeouts</name>
    <value>2</value>
  </property>

  <property>
    <name>dfs.socket.timeout</name>
    <value>5000</value>
  </property>

  <property>
    <name>dfs.socket.write.timeout</name>
    <value>5000</value>
  </property>

  <property>
    <name>dfs.domain.socket.path</name>
    <value>/var/lib/hadoop/dn_socket._PORT</value>
  </property>

  <property>
    <name>dfs.block.local-path-access.user</name>
    <value>hbase</value>
  </property>

  <property>
     <name>dfs.client.read.shortcircuit.skip.checksum</name>
     <value>true</value>
   </property>

  <property>
    <name>dfs.client.file-block-storage-locations.timeout</name>
    <value>3000</value>
  </property>

</configuration>
{code}

yarn-site.xml:
{code}
<?xml version="1.0"?>
<configuration>
  <property>
    <name>yarn.nodemanager.aux-services</name>
    <value>mapreduce.shuffle</value>
  </property>
  <property>
    <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
    <value>org.apache.hadoop.mapred.ShuffleHandler</value>
  </property>
  <property>
    <name>yarn.resourcemanager.address</name>
    <value>${MASTER_HOSTNAME}:10040</value>
    <description>In Server specified the port that Resource Manager will runn on. In client is used for connecting to Resource Manager</description>
  </property>
  <property>
    <name>yarn.resourcemanager.resource-tracker.address</name>
    <value>${MASTER_HOSTNAME}:8025</value>
    <description>Utilized by Node Manager for communication with Resource Manager</description>
  </property>
  <property>
    <name>yarn.resourcemanager.scheduler.address</name>
    <value>${MASTER_HOSTNAME}:8030</value>
     <description>Utilized by Application Masters to communicate with Resource Manager; in our case for MRAppMaster (MapReduce Application Master) to communicate with Resource Manager</description>
  </property>
  <property>
    <name>yarn.resourcemanager.admin.address</name>
    <value>${MASTER_HOSTNAME}:8141</value>
     <description>Utilized by administrative clients ($yarn rmadmin) to communicate with Resource Manager</description>
  </property>

  <property>
     <name>yarn.nodemanager.local-dirs</name>
     <value>${hadoop.data.dir0}/mapred/nodemanager,${hadoop.data.dir1}/mapred/nodemanager,${hadoop.data.dir2}/mapred/nodemanager,${hadoop.data.dir3}/mapred/nodemanager,${hadoop.data.dir4}/mapred/nodemanager,${hadoop.data.dir5}/mapred/nodemanager</value>
     <final>true</final>
     <description>Comma separated list of directories, where local data is persisted by Node Manager</description>
  </property>

  <property>
     <name>yarn.nodemanager.aux-services</name>
     <value>mapreduce.shuffle</value>
     <description>Long running service which executes on Node Manager(s) and provides MapReduce Sort and Shuffle functionality.</description>
  </property>

  <property>
     <name>yarn.log-aggregation-enable</name>
     <value>true</value>
    <description>Enable log aggregation so application logs are moved onto hdfs and are viewable via web ui after the application completed. The default location on hdfs is '/log' and can be changed via yarn.nodemanager.remote-app-log-dir property</description>
  </property>

 <property>
     <name>hadoop.security.authorization</name>
     <value>false</value>
     <description>Disable authorization fo development and clusters that do not require security</description>
   </property>

  <property>
    <description>Amount of physical memory, in MB, that can be allocated for containers.</description>
    <name>yarn.nodemanager.resource.memory-mb</name>
    <value>9000</value>
  </property>
</configuration>
{code}

hbase-site.xml:
{code}
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<configuration>
  <property>
    <name>hbase.zookeeper.quorum</name>
    <value>${MASTER_HOSTNAME}</value>
    <description>The directory shared by RegionServers.
    </description>
  </property>

  <property>
    <name>hbase.rootdir</name>
    <value>hdfs://${MASTER_HOSTNAME}:8020/hbase</value>
  </property>

  <property>
    <name>hbase.cluster.distributed</name>
    <value>true</value>
  </property>

  <property>
    <name>dfs.client.read.shortcircuit</name>
    <value>true</value>
  </property>

  <property>
    <name>dfs.block.local-path-access.user</name>
    <value>hbase</value>
  </property>

  <property>
    <name>dfs.domain.socket.path</name>
    <value>/var/lib/hadoop/dn_socket._PORT</value>
  </property>

  <property>
     <name>dfs.client.read.shortcircuit.skip.checksum</name>
     <value>true</value>
   </property>

   <property>
     <name>hbase.IntegrationTestDataIngestSlowDeterministic.runtime</name>
     <value>3600000</value>
   </property>
</configuration>
{code}
                
> Test Big Linked List fails on Hadoop 2.1.0
> ------------------------------------------
>
>                 Key: HBASE-9338
>                 URL: https://issues.apache.org/jira/browse/HBASE-9338
>             Project: HBase
>          Issue Type: Bug
>          Components: test
>    Affects Versions: 0.96.0
>            Reporter: Elliott Clark
>            Assignee: Elliott Clark
>            Priority: Critical
>             Fix For: 0.96.0
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira