You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Andrew Onischuk (JIRA)" <ji...@apache.org> on 2015/04/27 16:53:38 UTC

[jira] [Created] (AMBARI-10764) Incorrect configuration of spark-defaults.conf

Andrew Onischuk created AMBARI-10764:
----------------------------------------

             Summary: Incorrect configuration of spark-defaults.conf
                 Key: AMBARI-10764
                 URL: https://issues.apache.org/jira/browse/AMBARI-10764
             Project: Ambari
          Issue Type: Bug
            Reporter: Andrew Onischuk
            Assignee: Andrew Onischuk
             Fix For: 2.1.0


Due to configuration issue in spark-defaults.conf, All Spark applications
fails to start containers.

    
    
    
    Stack trace: ExitCodeException exitCode=1: /grid/0/hadoop/yarn/local/usercache/hrt_qa/appcache/application_1429516150624_0124/container_1429516150624_0124_02_000003/launch_container.sh: line 14: $PWD:$PWD/__spark__.jar:$HADOOP_CONF_DIR:/usr/hdp/current/hadoop-client/*:/usr/hdp/current/hadoop-client/lib/*:/usr/hdp/current/hadoop-hdfs-client/*:/usr/hdp/current/hadoop-hdfs-client/lib/*:/usr/hdp/current/hadoop-yarn-client/*:/usr/hdp/current/hadoop-yarn-client/lib/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar:/etc/hadoop/conf/secure:$PWD/__app__.jar:$PWD/*: bad substitution
    
            at org.apache.hadoop.util.Shell.runCommand(Shell.java:545)
            at org.apache.hadoop.util.Shell.run(Shell.java:456)
            at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:722)
            at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:211)
            at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302)
            at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)
            at java.util.concurrent.FutureTask.run(FutureTask.java:262)
            at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
            at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
            at java.lang.Thread.run(Thread.java:745)
    
    
    Container exited with a non-zero exit code 1
    

Issues with spark-defaults.conf

  * Does not set values for spark.driver.extraJavaOptions and spark.yarn.am.extraJavaOptions 

**correct config value**
    
        
    spark.yarn.am.extraJavaOptions    -Dhdp.version=2.3.0.0-1644 
    spark.driver.extraJavaOptions     -Dhdp.version=2.3.0.0-1644 
    

  * spark.yarn.historyServer.address property is not set 

**correct config value**
    
        
    spark.yarn.historyServer.address         os-amb-r6-us-1429252813-spark-2.novalocal:18080
    

  * new spark config does not set spark.yarn.max_executor.failures and spark.yarn.services property. Is it expected? zzhang can you please confirm this?

Attaching current and expected Spark-defaults.conf.





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)