You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by sr...@apache.org on 2016/03/21 08:59:29 UTC

spark git commit: [MINOR][DOCS] Add proper periods and spaces for CLI help messages and `config` doc.

Repository: spark
Updated Branches:
  refs/heads/master 20fd25410 -> 761c2d1b6


[MINOR][DOCS] Add proper periods and spaces for CLI help messages and `config` doc.

## What changes were proposed in this pull request?

This PR adds some proper periods and spaces to Spark CLI help messages and SQL/YARN conf docs for consistency.

## How was this patch tested?

Manual.

Author: Dongjoon Hyun <do...@apache.org>

Closes #11848 from dongjoon-hyun/add_proper_period_and_space.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/761c2d1b
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/761c2d1b
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/761c2d1b

Branch: refs/heads/master
Commit: 761c2d1b6ee0482e2ba15664c4938eb121dda070
Parents: 20fd254
Author: Dongjoon Hyun <do...@apache.org>
Authored: Mon Mar 21 08:00:09 2016 +0000
Committer: Sean Owen <so...@cloudera.com>
Committed: Mon Mar 21 08:00:09 2016 +0000

----------------------------------------------------------------------
 .../spark/deploy/SparkSubmitArguments.scala     |  6 ++---
 .../org/apache/spark/sql/internal/SQLConf.scala | 24 ++++++++++----------
 .../org/apache/spark/deploy/yarn/config.scala   |  2 +-
 3 files changed, 16 insertions(+), 16 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/761c2d1b/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala b/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
index a62096d..ec6d484 100644
--- a/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
@@ -524,9 +524,9 @@ private[deploy] class SparkSubmitArguments(args: Seq[String], env: Map[String, S
         |  --proxy-user NAME           User to impersonate when submitting the application.
         |                              This argument does not work with --principal / --keytab.
         |
-        |  --help, -h                  Show this help message and exit
-        |  --verbose, -v               Print additional debug output
-        |  --version,                  Print the version of current Spark
+        |  --help, -h                  Show this help message and exit.
+        |  --verbose, -v               Print additional debug output.
+        |  --version,                  Print the version of current Spark.
         |
         | Spark standalone with cluster deploy mode only:
         |  --driver-cores NUM          Cores for driver (Default: 1).

http://git-wip-us.apache.org/repos/asf/spark/blob/761c2d1b/sql/core/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala b/sql/core/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
index 473cde5..9e0878a 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
@@ -210,11 +210,11 @@ object SQLConf {
 
   val ALLOW_MULTIPLE_CONTEXTS = booleanConf("spark.sql.allowMultipleContexts",
     defaultValue = Some(true),
-    doc = "When set to true, creating multiple SQLContexts/HiveContexts is allowed." +
+    doc = "When set to true, creating multiple SQLContexts/HiveContexts is allowed. " +
       "When set to false, only one SQLContext/HiveContext is allowed to be created " +
       "through the constructor (new SQLContexts/HiveContexts created through newSession " +
-      "method is allowed). Please note that this conf needs to be set in Spark Conf. Once" +
-      "a SQLContext/HiveContext has been created, changing the value of this conf will not" +
+      "method is allowed). Please note that this conf needs to be set in Spark Conf. Once " +
+      "a SQLContext/HiveContext has been created, changing the value of this conf will not " +
       "have effect.",
     isPublic = true)
 
@@ -238,7 +238,7 @@ object SQLConf {
 
   val PREFER_SORTMERGEJOIN = booleanConf("spark.sql.join.preferSortMergeJoin",
       defaultValue = Some(true),
-      doc = "When true, prefer sort merge join over shuffle hash join",
+      doc = "When true, prefer sort merge join over shuffle hash join.",
       isPublic = false)
 
   val AUTO_BROADCASTJOIN_THRESHOLD = intConf("spark.sql.autoBroadcastJoinThreshold",
@@ -252,8 +252,8 @@ object SQLConf {
     "spark.sql.defaultSizeInBytes",
     doc = "The default table size used in query planning. By default, it is set to a larger " +
       "value than `spark.sql.autoBroadcastJoinThreshold` to be more conservative. That is to say " +
-      "by default the optimizer will not choose to broadcast a table unless it knows for sure its" +
-      "size is small enough.",
+      "by default the optimizer will not choose to broadcast a table unless it knows for sure " +
+      "its size is small enough.",
     isPublic = false)
 
   val SHUFFLE_PARTITIONS = intConf("spark.sql.shuffle.partitions",
@@ -275,7 +275,7 @@ object SQLConf {
       doc = "The advisory minimal number of post-shuffle partitions provided to " +
         "ExchangeCoordinator. This setting is used in our test to make sure we " +
         "have enough parallelism to expose issues that will not be exposed with a " +
-        "single partition. When the value is a non-positive value, this setting will" +
+        "single partition. When the value is a non-positive value, this setting will " +
         "not be provided to ExchangeCoordinator.",
       isPublic = false)
 
@@ -391,7 +391,7 @@ object SQLConf {
 
   // This is only used for the thriftserver
   val THRIFTSERVER_POOL = stringConf("spark.sql.thriftserver.scheduler.pool",
-    doc = "Set a Fair Scheduler pool for a JDBC client session")
+    doc = "Set a Fair Scheduler pool for a JDBC client session.")
 
   val THRIFTSERVER_UI_STATEMENT_LIMIT = intConf("spark.sql.thriftserver.ui.retainedStatements",
     defaultValue = Some(200),
@@ -433,7 +433,7 @@ object SQLConf {
 
   val BUCKETING_ENABLED = booleanConf("spark.sql.sources.bucketing.enabled",
     defaultValue = Some(true),
-    doc = "When false, we will treat bucketed table as normal table")
+    doc = "When false, we will treat bucketed table as normal table.")
 
   // The output committer class used by HadoopFsRelation. The specified class needs to be a
   // subclass of org.apache.hadoop.mapreduce.OutputCommitter.
@@ -482,7 +482,7 @@ object SQLConf {
   val RUN_SQL_ON_FILES = booleanConf("spark.sql.runSQLOnFiles",
     defaultValue = Some(true),
     isPublic = false,
-    doc = "When true, we could use `datasource`.`path` as table in SQL query"
+    doc = "When true, we could use `datasource`.`path` as table in SQL query."
   )
 
   val PARSER_SUPPORT_QUOTEDID = booleanConf("spark.sql.parser.supportQuotedIdentifiers",
@@ -501,7 +501,7 @@ object SQLConf {
   val WHOLESTAGE_CODEGEN_ENABLED = booleanConf("spark.sql.codegen.wholeStage",
     defaultValue = Some(true),
     doc = "When true, the whole stage (of multiple operators) will be compiled into single java" +
-      " method",
+      " method.",
     isPublic = false)
 
   val FILES_MAX_PARTITION_BYTES = longConf("spark.sql.files.maxPartitionBytes",
@@ -511,7 +511,7 @@ object SQLConf {
 
   val EXCHANGE_REUSE_ENABLED = booleanConf("spark.sql.exchange.reuse",
     defaultValue = Some(true),
-    doc = "When true, the planner will try to find out duplicated exchanges and re-use them",
+    doc = "When true, the planner will try to find out duplicated exchanges and re-use them.",
     isPublic = false)
 
   object Deprecated {

http://git-wip-us.apache.org/repos/asf/spark/blob/761c2d1b/yarn/src/main/scala/org/apache/spark/deploy/yarn/config.scala
----------------------------------------------------------------------
diff --git a/yarn/src/main/scala/org/apache/spark/deploy/yarn/config.scala b/yarn/src/main/scala/org/apache/spark/deploy/yarn/config.scala
index 10cd6d0..0789567 100644
--- a/yarn/src/main/scala/org/apache/spark/deploy/yarn/config.scala
+++ b/yarn/src/main/scala/org/apache/spark/deploy/yarn/config.scala
@@ -104,7 +104,7 @@ package object config {
   /* Cluster-mode launcher configuration. */
 
   private[spark] val WAIT_FOR_APP_COMPLETION = ConfigBuilder("spark.yarn.submit.waitAppCompletion")
-    .doc("In cluster mode, whether to wait for the application to finishe before exiting the " +
+    .doc("In cluster mode, whether to wait for the application to finish before exiting the " +
       "launcher process.")
     .booleanConf
     .withDefault(true)


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org