You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by sr...@apache.org on 2018/12/03 16:02:28 UTC
spark git commit: [SPARK-26235][CORE] Change log level for
ClassNotFoundException/NoClassDefFoundError in SparkSubmit to Error
Repository: spark
Updated Branches:
refs/heads/master 8534d753e -> 6e4e70fe7
[SPARK-26235][CORE] Change log level for ClassNotFoundException/NoClassDefFoundError in SparkSubmit to Error
## What changes were proposed in this pull request?
In my local setup, I set log4j root category as ERROR (https://stackoverflow.com/questions/27781187/how-to-stop-info-messages-displaying-on-spark-console , first item show up if we google search "set spark log level".) When I run such command
```
spark-submit --class foo bar.jar
```
Nothing shows up, and the script exits.
After quick investigation, I think the log level for ClassNotFoundException/NoClassDefFoundError in SparkSubmit should be ERROR instead of WARN. Since the whole process exit because of the exception/error.
Before https://github.com/apache/spark/pull/20925, the message is not controlled by `log4j.rootCategory`.
## How was this patch tested?
Manual check.
Closes #23189 from gengliangwang/changeLogLevel.
Authored-by: Gengliang Wang <ge...@databricks.com>
Signed-off-by: Sean Owen <se...@databricks.com>
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6e4e70fe
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6e4e70fe
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6e4e70fe
Branch: refs/heads/master
Commit: 6e4e70fe7bc3e103b8538748511261bb43cf3548
Parents: 8534d75
Author: Gengliang Wang <ge...@databricks.com>
Authored: Mon Dec 3 10:02:15 2018 -0600
Committer: Sean Owen <se...@databricks.com>
Committed: Mon Dec 3 10:02:15 2018 -0600
----------------------------------------------------------------------
.../src/main/scala/org/apache/spark/deploy/SparkSubmit.scala | 8 ++++++--
1 file changed, 6 insertions(+), 2 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/6e4e70fe/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala b/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala
index 324f6f8..d4055cb 100644
--- a/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala
@@ -813,14 +813,14 @@ private[spark] class SparkSubmit extends Logging {
mainClass = Utils.classForName(childMainClass)
} catch {
case e: ClassNotFoundException =>
- logWarning(s"Failed to load $childMainClass.", e)
+ logError(s"Failed to load class $childMainClass.")
if (childMainClass.contains("thriftserver")) {
logInfo(s"Failed to load main class $childMainClass.")
logInfo("You need to build Spark with -Phive and -Phive-thriftserver.")
}
throw new SparkUserAppException(CLASS_NOT_FOUND_EXIT_STATUS)
case e: NoClassDefFoundError =>
- logWarning(s"Failed to load $childMainClass: ${e.getMessage()}")
+ logError(s"Failed to load $childMainClass: ${e.getMessage()}")
if (e.getMessage.contains("org/apache/hadoop/hive")) {
logInfo(s"Failed to load hive class.")
logInfo("You need to build Spark with -Phive and -Phive-thriftserver.")
@@ -915,6 +915,8 @@ object SparkSubmit extends CommandLineUtils with Logging {
override protected def logInfo(msg: => String): Unit = self.logInfo(msg)
override protected def logWarning(msg: => String): Unit = self.logWarning(msg)
+
+ override protected def logError(msg: => String): Unit = self.logError(msg)
}
}
@@ -922,6 +924,8 @@ object SparkSubmit extends CommandLineUtils with Logging {
override protected def logWarning(msg: => String): Unit = printMessage(s"Warning: $msg")
+ override protected def logError(msg: => String): Unit = printMessage(s"Error: $msg")
+
override def doSubmit(args: Array[String]): Unit = {
try {
super.doSubmit(args)
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org