You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by GitBox <gi...@apache.org> on 2020/02/19 21:46:05 UTC

[GitHub] [spark] gengliangwang commented on a change in pull request #27639: [SPARK-30881][SQL][Doc]Revise the doc of spark.sql.sources.parallelPartitionDiscovery.threshold

gengliangwang commented on a change in pull request #27639: [SPARK-30881][SQL][Doc]Revise the doc of spark.sql.sources.parallelPartitionDiscovery.threshold
URL: https://github.com/apache/spark/pull/27639#discussion_r381563401
 
 

 ##########
 File path: sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
 ##########
 @@ -878,8 +878,8 @@ object SQLConf {
     buildConf("spark.sql.sources.parallelPartitionDiscovery.threshold")
       .doc("The maximum number of paths allowed for listing files at driver side. If the number " +
         "of detected paths exceeds this value during partition discovery, it tries to list the " +
-        "files with another Spark distributed job. This applies to Parquet, ORC, CSV, JSON and " +
-        "LibSVM data sources.")
+        "files with another Spark distributed job. This configuration is effective only when " +
+        "using file-based sources such as Parquet, JSON and ORC.")
 
 Review comment:
   This follows the doc of `spark.sql.files.maxPartitionBytes` and `spark.sql.files.openCostInBytes`

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org