You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@whirr.apache.org by "Andrei Savu (Commented) (JIRA)" <ji...@apache.org> on 2011/12/06 18:27:40 UTC

[jira] [Commented] (WHIRR-441) Precondition failure: IndexOutOfBoundsException on cluster setup

    [ https://issues.apache.org/jira/browse/WHIRR-441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13163706#comment-13163706 ] 

Andrei Savu commented on WHIRR-441:
-----------------------------------

Thanks Evan for reporting this. I think you should use options like whirr.instance-templates-minimum-number-of-instances, whirr.max-startup-retries to improve your odds of starting larger size clusters. 

See the configuration guide:
http://whirr.apache.org/docs/0.6.0/configuration-guide.html
                
> Precondition failure:  IndexOutOfBoundsException on cluster setup
> -----------------------------------------------------------------
>
>                 Key: WHIRR-441
>                 URL: https://issues.apache.org/jira/browse/WHIRR-441
>             Project: Whirr
>          Issue Type: Bug
>    Affects Versions: 0.6.0
>         Environment: 64 bit Amazon linux AMI w/ Cloudera CDH3U2 hadoop/hive stack
>            Reporter: Evan Pollan
>
> I was spinning up a 16 node cluster this morning, and, after a series of errors (not uncommon) there was a precondition assertion failure that left the whirr JVM running, but dormant for about 20 minutes.  I haven't seen this before using the same cluster config and whirr version, and I'm trying again to see if it's reproducible.
> Here's the error:
> Starting 15 node(s) with roles [hadoop-datanode, hadoop-tasktracker]
> Starting 1 node(s) with roles [hadoop-jobtracker, hadoop-namenode]
> << problem applying options to node(us-east-1/sir-b61d7212): 
> org.jclouds.aws.AWSResponseException: request POST https://ec2.us-east-1.amazonaws.com/ HTTP/1.1 failed with code 400, error: AWSError{requestId='9530b126-fae6-43c8-86a4-b7e2a865c8a1', requestToken='null', code='InternalError', message='An internal error has occurred', context='{Response=, Errors=}'}
> 	at org.jclouds.aws.handlers.ParseAWSErrorFromXmlContent.handleError(ParseAWSErrorFromXmlContent.java:74)
> 	at org.jclouds.http.handlers.DelegatingErrorHandler.handleError(DelegatingErrorHandler.java:69)
> 	at org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.shouldContinue(BaseHttpCommandExecutorService.java:200)
> 	at org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.call(BaseHttpCommandExecutorService.java:165)
> 	at org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.call(BaseHttpCommandExecutorService.java:134)
> 	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
> 	at java.lang.Thread.run(Thread.java:636)
> << problem applying options to node(us-east-1/sir-d4907012): 
> org.jclouds.aws.AWSResponseException: request POST https://ec2.us-east-1.amazonaws.com/ HTTP/1.1 failed with code 400, error: AWSError{requestId='69215a6b-5455-402f-ae0c-aaaca6245cb6', requestToken='null', code='InternalError', message='An internal error has occurred', context='{Response=, Errors=}'}
> 	at org.jclouds.aws.handlers.ParseAWSErrorFromXmlContent.handleError(ParseAWSErrorFromXmlContent.java:74)
> 	at org.jclouds.http.handlers.DelegatingErrorHandler.handleError(DelegatingErrorHandler.java:69)
> 	at org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.shouldContinue(BaseHttpCommandExecutorService.java:200)
> 	at org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.call(BaseHttpCommandExecutorService.java:165)
> 	at org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.call(BaseHttpCommandExecutorService.java:134)
> 	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
> 	at java.lang.Thread.run(Thread.java:636)
> Nodes started: [[id=us-east-1/i-6dcec30e, providerId=i-6dcec30e, group=logs-cluster, name=null, location=[id=us-east-1c, scope=ZONE, description=us-east-1c, parent=us-east-1, iso3166Codes=[US-VA], metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null, family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true, description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml], state=RUNNING, loginPort=22, hostname=ip-10-196-130-159, privateAddresses=[10.196.130.159], publicAddresses=[204.236.240.255], hardware=[id=m1.xlarge, providerId=m1.xlarge, name=null, processors=[[cores=4.0, speed=2.0]], ram=15360, volumes=[[id=null, type=LOCAL, size=10.0, device=/dev/sda1, durable=false, isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sdc, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sdd, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sde, durable=false, isBootDevice=false]], supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()), tags=[]], loginUser=ubuntu, userMetadata={}, tags=[]]]
> Starting 2 node(s) with roles [hadoop-datanode, hadoop-tasktracker]
> Dying because - net.schmizz.sshj.transport.TransportException: Broken transport; encountered EOF
> Dying because - net.schmizz.sshj.transport.TransportException: Broken transport; encountered EOF
> <<kex done>> woke to: net.schmizz.sshj.transport.TransportException: Broken transport; encountered EOF
> << (ubuntu@50.16.99.93:22) error acquiring SSHClient(ubuntu@50.16.99.93:22): Broken transport; encountered EOF
> net.schmizz.sshj.transport.TransportException: Broken transport; encountered EOF
> 	at net.schmizz.sshj.transport.Reader.run(Reader.java:70)
> Nodes started: [[id=us-east-1/i-cbc6cba8, providerId=i-cbc6cba8, group=logs-cluster, name=null, location=[id=us-east-1d, scope=ZONE, description=us-east-1d, parent=us-east-1, iso3166Codes=[US-VA], metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null, family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true, description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml], state=RUNNING, loginPort=22, hostname=ip-10-120-239-36, privateAddresses=[10.120.239.36], publicAddresses=[50.16.99.93], hardware=[id=m1.xlarge, providerId=m1.xlarge, name=null, processors=[[cores=4.0, speed=2.0]], ram=15360, volumes=[[id=null, type=LOCAL, size=10.0, device=/dev/sda1, durable=false, isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sdc, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sdd, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sde, durable=false, isBootDevice=false]], supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()), tags=[]], loginUser=ubuntu, userMetadata={}, tags=[]], [id=us-east-1/i-cdc6cbae, providerId=i-cdc6cbae, group=logs-cluster, name=null, location=[id=us-east-1d, scope=ZONE, description=us-east-1d, parent=us-east-1, iso3166Codes=[US-VA], metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null, family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true, description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml], state=RUNNING, loginPort=22, hostname=ip-10-123-69-36, privateAddresses=[10.123.69.36], publicAddresses=[107.22.69.237], hardware=[id=m1.xlarge, providerId=m1.xlarge, name=null, processors=[[cores=4.0, speed=2.0]], ram=15360, volumes=[[id=null, type=LOCAL, size=10.0, device=/dev/sda1, durable=false, isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sdc, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sdd, durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0, device=/dev/sde, durable=false, isBootDevice=false]], supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()), tags=[]], loginUser=ubuntu, userMetadata={}, tags=[]]]
> Deleting failed node node us-east-1/sir-d4907012
> Deleting failed node node us-east-1/sir-b61d7212
> Deleting failed node node us-east-1/i-23cec340
> Node deleted: us-east-1/sir-b61d7212
> Node deleted: us-east-1/sir-d4907012
> Node deleted: us-east-1/i-23cec340
> Exception in thread "main" java.lang.IndexOutOfBoundsException: index (0) must be less than size (0)
> 	at com.google.common.base.Preconditions.checkElementIndex(Preconditions.java:301)
> 	at com.google.common.base.Preconditions.checkElementIndex(Preconditions.java:280)
> 	at com.google.common.collect.Iterables.get(Iterables.java:649)
> 	at org.apache.whirr.actions.BootstrapClusterAction$1.apply(BootstrapClusterAction.java:226)
> 	at org.apache.whirr.actions.BootstrapClusterAction$1.apply(BootstrapClusterAction.java:223)
> 	at com.google.common.collect.Iterators$8.next(Iterators.java:765)
> 	at java.util.AbstractCollection.addAll(AbstractCollection.java:322)
> 	at java.util.LinkedHashSet.<init>(LinkedHashSet.java:169)
> 	at com.google.common.collect.Sets.newLinkedHashSet(Sets.java:264)
> 	at org.apache.whirr.actions.BootstrapClusterAction.getInstances(BootstrapClusterAction.java:222)
> 	at org.apache.whirr.actions.BootstrapClusterAction.doAction(BootstrapClusterAction.java:141)
> 	at org.apache.whirr.actions.ScriptBasedClusterAction.execute(ScriptBasedClusterAction.java:80)
> 	at org.apache.whirr.ClusterController.launchCluster(ClusterController.java:106)
> 	at org.apache.whirr.cli.command.LaunchClusterCommand.run(LaunchClusterCommand.java:62)
> 	at org.apache.whirr.cli.Main.run(Main.java:64)
> 	at org.apache.whirr.cli.Main.main(Main.java:97)
> Here's the config file:
> =======================
> whirr.instance-templates=1 hadoop-jobtracker+hadoop-namenode,15 hadoop-datanode+hadoop-tasktracker
> whirr.hadoop.install-function=install_cdh_hadoop
> whirr.hadoop.configure-function=configure_cdh_hadoop
> whirr.provider=aws-ec2
> whirr.cluster-name=logs-cluster
> whirr.identity=ACCESS KEY ID GOES HERE
> whirr.credential=SECRET ACCESS KEY GOES HERE
> whirr.private-key-file=${sys:user.home}/.ssh/id_rsa
> whirr.public-key-file=${sys:user.home}/.ssh/id_rsa.pub
> whirr.hardware-id=m1.xlarge
> # Using 64 bit Ubuntu 10.04 to avoid defect https://issues.apache.org/jira/browse/WHIRR-148
> whirr.image-id=us-east-1/ami-da0cf8b3
> whirr.aws-ec2-spot-price=1.00

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira