You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Edward Capriolo (JIRA)" <ji...@apache.org> on 2014/04/13 18:08:15 UTC

[jira] [Resolved] (HIVE-6212) Using Presto-0.56 for sql query,but HiveServer the console print java.lang.OutOfMemoryError: Java heap space

     [ https://issues.apache.org/jira/browse/HIVE-6212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Edward Capriolo resolved HIVE-6212.
-----------------------------------

    Resolution: Won't Fix

Contact presto developers. 

> Using Presto-0.56 for sql query,but HiveServer the console print java.lang.OutOfMemoryError: Java heap space
> ------------------------------------------------------------------------------------------------------------
>
>                 Key: HIVE-6212
>                 URL: https://issues.apache.org/jira/browse/HIVE-6212
>             Project: Hive
>          Issue Type: Bug
>          Components: HiveServer2
>    Affects Versions: 0.11.0
>         Environment: HADOOP ENVIRONMENT IS CDH5+CDH5-HIVE-0.11+PRESTO-0.56
>            Reporter: apachehadoop
>             Fix For: 0.11.0
>
>
> Hi friends:
> Now I can't open the page https://groups.google.com/forum/#!forum/presto-users ,so show my question here.
> I have started hiveserver and started presto-server on a machine with commands below:
> hive --service hiveserver -p 9083
> ./launcher run
> When I use the presto-client-cli command ./presto --server localhost:9083 --catalog hive --schema default ,the console shows presto:default>,input the command as show tables the console prints Error running command: java.nio.channels.ClosedChannelException,
> and the hiveserver console print as below:
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> Exception in thread "pool-1-thread-1" java.lang.OutOfMemoryError: Java heap space
> at org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:353)
> at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:215)
> at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:27)
> at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
> at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
> at java.lang.Thread.run(Thread.java:662)
> my configuration file below:
> node.properties
> node.environment=production
> node.id=cc4a1bbf-5b98-4935-9fde-2cf1c98e8774
> node.data-dir=/home/hadoop/cloudera-5.0.0/presto-0.56/presto/data
> config.properties
> coordinator=true
> datasources=jmx
> http-server.http.port=8080
> presto-metastore.db.type=h2
> presto-metastore.db.filename=/home/hadoop/cloudera-5.0.0/presto-0.56/presto/db/MetaStore
> task.max-memory=1GB
> discovery-server.enabled=true
> discovery.uri=http://slave4:8080
> jvm.config
> -server
> -Xmx16G
> -XX:+UseConcMarkSweepGC
> -XX:+ExplicitGCInvokesConcurrent
> -XX:+CMSClassUnloadingEnabled
> -XX:+AggressiveOpts
> -XX:+HeapDumpOnOutOfMemoryError
> -XX:OnOutOfMemoryError=kill -9 %p
> -XX:PermSize=150M
> -XX:MaxPermSize=150M
> -XX:ReservedCodeCacheSize=150M
> -Xbootclasspath/p:/home/hadoop/cloudera-5.0.0/presto-0.56/presto-server-0.56/lib/floatingdecimal-0.1.jar
> log.properties
> com.facebook.presto=DEBUG
> catalog/hive.properties
> connector.name=hive-cdh4
> hive.metastore.uri=thrift://slave4:9083
> HADOOP ENVIRONMENT IS CDH5+CDH5-HIVE-0.11+PRESTO-0.56
> Last I had increased the Java heap size for the Hive metastore,but it still given me the same error informations ,please help me to check if that is a bug of CDH5.Now I have no idea,god !
> please help me ,thanks.
> **********************************************************************************************************************************************************************
> ========================================================================================================
> **********************************************************************************************************************************************************************
> Add some informations as below:
> Help,help,help!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
> I have test prest-server-0.55 and 0.56 and 0.57 on CDH4 +hive-0.10 or hive-0.11,but it still shown error informations above.
> ON coordinator machine the directory etc and configuration files as below:
> =================coordinator================================ 
>  ----------------config.properties:
> coordinator=true
> datasources=jmx
> http-server.http.port=8080
> presto-metastore.db.type=h2
> presto-metastore.db.filename=/home/hadoop/cloudera-5.0.0/presto-0.55/presto/db/MetaStore
> task.max-memory=1GB
> discovery-server.enabled=true
> discovery.uri=http://name:8080
> --------------jvm.config:
> -server
> -Xmx4G
> -XX:+UseConcMarkSweepGC
> -XX:+ExplicitGCInvokesConcurrent
> -XX:+CMSClassUnloadingEnabled
> -XX:+AggressiveOpts
> -XX:+HeapDumpOnOutOfMemoryError
> -XX:OnOutOfMemoryError=kill -9 %p
> -XX:PermSize=150M
> -XX:MaxPermSize=150M
> -XX:ReservedCodeCacheSize=150M
> -Xbootclasspath/p:/home/hadoop/cloudera-5.0.0/presto-0.55/presto-server-0.55/lib/floatingdecimal-0.1.jar
> -----------------log.properties:
> com.facebook.presto=DEBUG
> -----------------node.properties:
> node.environment=production
> #node.id=0699bf8f-ac4e-48f4-92a9-28c0d8862923
> node.id=name
> node.data-dir=/home/hadoop/cloudera-5.0.0/presto-0.55/presto/data
> ---------------hive.properties:
> connector.name=hive-cdh4
> hive.metastore.uri=thrift://name:10000
> ---------------jmx.properties:
> connector.name=jmx
> =================coordinator================================ 
> =================woker==================================== 
> ----------------config.properties:
> coordinator=false
> datasources=jmx,hive
> http-server.http.port=8080
> presto-metastore.db.type=h2
> presto-metastore.db.filename=/home/hadoop/cloudera-5.0.0/presto-0.55/presto/db/MetaStore
> task.max-memory=1GB
> discovery-server.enabled=true
> discovery.uri=http://name:8080
> --------------jvm.config:
> -server
> -Xmx4G
> -XX:+UseConcMarkSweepGC
> -XX:+ExplicitGCInvokesConcurrent
> -XX:+CMSClassUnloadingEnabled
> -XX:+AggressiveOpts
> -XX:+HeapDumpOnOutOfMemoryError
> -XX:OnOutOfMemoryError=kill -9 %p
> -XX:PermSize=150M
> -XX:MaxPermSize=150M
> -XX:ReservedCodeCacheSize=150M
> -Xbootclasspath/p:/home/hadoop/cloudera-5.0.0/presto-0.55/presto-server-0.55/lib/floatingdecimal-0.1.jar
> -----------------log.properties:
> com.facebook.presto=DEBUG
> -----------------node.properties:
> node.environment=production
> #node.id=773a6b4e-9fd0-4342-8d96-59d1f58ac7cd
> node.id=data1
> node.data-dir=/home/hadoop/cloudera-5.0.0/presto-0.55/presto/data
> ---------------hive.properties:
> connector.name=hive-cdh4
> hive.metastore.uri=thrift://name:10000
> ---------------jmx.properties:
> connector.name=jmx
> =================woker====================================
> =================discovery-server===========================
> --------------config.properties:
> http-server.http.port=8411
> discovery.uri=http://name:8080
> --------------node.properties:
> node.environment=production
> #node.id=b45edeee-4870-420f-919b-9f8487b9750e
> #node.id=0699bf8f-ac4e-48f4-92a9-28c0d8862923
> node.id=name
> node.data-dir=/home/hadoop/cloudera-5.0.0/presto-discovery-data
> ----------------jvm.config
> -server
> -Xmx1G
> -XX:+UseConcMarkSweepGC
> -XX:+ExplicitGCInvokesConcurrent
> -XX:+AggressiveOpts
> -XX:+HeapDumpOnOutOfMemoryError
> -XX:OnOutOfMemoryError=kill -9 %p
> =================discovery-server===========================
> I had started hiveserver and coordinator service and discovery-server on name machine,worker on data1 machine, all that ware Normal.When I started to use the command which was as below:
> ./presto --server name:10000 --catalog hive --schema default
> the console of hiveserver that shown :
> Exception in thread "pool-1-thread-1" java.lang.OutOfMemoryError: Java heap space
>         at org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:353)
>         at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:215)
>         at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:27)
>         at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
>         at java.lang.Thread.run(Thread.java:662)
> MeanWhile,I had increased the java heap size of hive to  2g or 4g,however it still throwed exceptions as above.
> Please give me some ideas because of you are Professional.
> Thanks.



--
This message was sent by Atlassian JIRA
(v6.2#6252)