You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@cassandra.apache.org by Michael Shuler <mi...@pbandjelly.org> on 2014/06/04 19:32:19 UTC

Re: Too Many Open Files (sockets) - VNodes - Map/Reduce Job

(this is probably a better question for the user list - cc/reply-to set)

Allow more files to be open  :)

http://www.datastax.com/documentation/cassandra/1.2/cassandra/install/installRecommendSettings.html

-- 
Kind regards,
Michael


On 06/04/2014 12:15 PM, Florian Dambrine wrote:
> Hi every body,
>
> We are running ElasticMapReduce Jobs from Amazon on a 25 nodes Cassandra
> cluster (with VNodes). Since we have increased the size of the cluster we
> are facing a too many open files (due to sockets) exception when creating
> the splits. Does anyone has an idea?
>
> Thanks,
>
> Here is the stacktrace:
>
>
> 14/06/04 03:23:24 INFO mapred.JobClient: Default number of map tasks: null
> 14/06/04 03:23:24 INFO mapred.JobClient: Setting default number of map
> tasks based on cluster size to : 80
> 14/06/04 03:23:24 INFO mapred.JobClient: Default number of reduce tasks: 26
> 14/06/04 03:23:25 INFO security.ShellBasedUnixGroupsMapping: add
> hadoop to shell userGroupsCache
> 14/06/04 03:23:25 INFO mapred.JobClient: Setting group to hadoop
> 14/06/04 03:23:41 ERROR transport.TSocket: Could not configure socket.
> java.net.SocketException: Too many open files
> 	at java.net.Socket.createImpl(Socket.java:447)
> 	at java.net.Socket.getImpl(Socket.java:510)
> 	at java.net.Socket.setSoLinger(Socket.java:984)
> 	at org.apache.thrift.transport.TSocket.initSocket(TSocket.java:118)
> 	at org.apache.thrift.transport.TSocket.<init>(TSocket.java:109)
> 	at org.apache.thrift.transport.TSocket.<init>(TSocket.java:94)
> 	at org.apache.cassandra.thrift.TFramedTransportFactory.openTransport(TFramedTransportFactory.java:39)
> 	at org.apache.cassandra.hadoop.ConfigHelper.createConnection(ConfigHelper.java:558)
> 	at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSubSplits(AbstractColumnFamilyInputFormat.java:286)
> 	at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.access$200(AbstractColumnFamilyInputFormat.java:61)
> 	at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat$SplitCallable.call(AbstractColumnFamilyInputFormat.java:236)
> 	at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat$SplitCallable.call(AbstractColumnFamilyInputFormat.java:221)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:744)
>
>
>