You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Zhenhua Wang (JIRA)" <ji...@apache.org> on 2015/10/29 07:19:28 UTC

[jira] [Created] (SPARK-11398) misleading dialect conf at the start of spark-sql

Zhenhua Wang created SPARK-11398:
------------------------------------

             Summary: misleading dialect conf at the start of spark-sql
                 Key: SPARK-11398
                 URL: https://issues.apache.org/jira/browse/SPARK-11398
             Project: Spark
          Issue Type: Bug
          Components: SQL
            Reporter: Zhenhua Wang
            Priority: Minor


When we start bin/spark-sql, the default context is HiveContext, and the corresponding dialect is hiveql.
However, if we type "set spark.sql.dialect;", the result is "sql", which is inconsistent with the actual dialect and is misleading. For example, we can create tables which is only allowed in hiveql, but this dialect conf shows it's "sql".
Although This problem will not cause any execution error, it's misleading to spark sql users. Therefore I think we should fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org