You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Szehon Ho (JIRA)" <ji...@apache.org> on 2015/04/10 02:44:12 UTC

[jira] [Created] (HIVE-10291) Hive on Spark job configuration needs to be logged [Spark Branch]

Szehon Ho created HIVE-10291:
--------------------------------

             Summary: Hive on Spark job configuration needs to be logged [Spark Branch]
                 Key: HIVE-10291
                 URL: https://issues.apache.org/jira/browse/HIVE-10291
             Project: Hive
          Issue Type: Sub-task
          Components: Spark
    Affects Versions: 1.1.0
            Reporter: Szehon Ho
            Assignee: Szehon Ho


In a Hive on MR job, all the job properties are put into the JobConf, which can then be viewed via the MR2 HistoryServer's Job UI.

However, in Hive on Spark we are submitting an application that is long-lived.  Hence, we only put properties into the SparkConf relevant to application submission (spark and yarn properties).  Only these are viewable through the Spark HistoryServer Application UI.

It is the Hive application code (RemoteDriver, aka RemoteSparkContext) that is responsible for serializing and deserializing the job.xml per job (ie, query) within the application.  Thus, for supportability we also need to give an equivalent mechanism to print the job.xml per job.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)