You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Idan Zalzberg (JIRA)" <ji...@apache.org> on 2015/01/19 10:54:34 UTC

[jira] [Created] (SPARK-5318) Add ability to control partition count in SparkSql

Idan Zalzberg created SPARK-5318:
------------------------------------

             Summary: Add ability to control partition count in SparkSql
                 Key: SPARK-5318
                 URL: https://issues.apache.org/jira/browse/SPARK-5318
             Project: Spark
          Issue Type: New Feature
          Components: SQL
            Reporter: Idan Zalzberg


When using SparkSql, e.g. sqlContext.sql("..."), spark might need to read hadoop files.
However, unlike the hadoopFile API, there is no documented way to set the minimal partition count when reading.
There is an undocumented way, though, using "mapred.map.tasks" in hiveConf

I suggest we make a documented way to do it, in the exact same way (possibly with a better name)




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org