You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Giambattista (JIRA)" <ji...@apache.org> on 2017/01/03 14:34:58 UTC

[jira] [Created] (SPARK-19059) Unable to retrieve data from a parquet table whose name starts with underscore

Giambattista created SPARK-19059:
------------------------------------

             Summary: Unable to retrieve data from a parquet table whose name starts with underscore
                 Key: SPARK-19059
                 URL: https://issues.apache.org/jira/browse/SPARK-19059
             Project: Spark
          Issue Type: Bug
          Components: Spark Core
    Affects Versions: 2.1.0
            Reporter: Giambattista
             Fix For: 2.1.1


It looks like there is some bug introduced in Spark 2.1.0 preventing to read data from a parquet table (hive support is enabled) whose name starts with underscore. CREATE and INSERT statements on the same table instead seems to work as expected.

The problem can be reproduced from spark-shell through the following steps:
1) Create a table with some values
scala> spark.sql("CREATE TABLE `_a`(i INT) USING parquet").show
scala> spark.sql("INSERT INTO `_a` VALUES (1), (2), (3)").show

2) Select data from the just created and filled table --> no results
scala> spark.sql("SELECT * FROM `_a`").show
+---+
|  i|
+---+
+---+

3) rename the table so that the prefixing underscore disappears
scala> spark.sql("ALTER TABLE `_a` RENAME TO `a`").show

4) select data from the just renamed table --> results are shown
scala> spark.sql("SELECT * FROM `a`").show
+---+
|  i|
+---+
|  1|
|  2|
|  3|
+---+







--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org