You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@ambari.apache.org by "Doroszlai, Attila (JIRA)" <ji...@apache.org> on 2017/07/28 14:52:00 UTC
[jira] [Updated] (AMBARI-21598) Spark Thrift Server stopped after
express upgrade due to undefined port
[ https://issues.apache.org/jira/browse/AMBARI-21598?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Doroszlai, Attila updated AMBARI-21598:
---------------------------------------
Status: Patch Available (was: In Progress)
> Spark Thrift Server stopped after express upgrade due to undefined port
> -----------------------------------------------------------------------
>
> Key: AMBARI-21598
> URL: https://issues.apache.org/jira/browse/AMBARI-21598
> Project: Ambari
> Issue Type: Bug
> Components: ambari-upgrade
> Affects Versions: 2.5.2
> Environment: Source Ambari Version:- 2.2.2
> Target Ambari Version:-:-ambari-2.5.2.0-189
> Source Stack:-BigInsights-4.2.0.0
> target Stack:- HDP-2.6.2.0-124
> Reporter: Pradarttana
> Assignee: Doroszlai, Attila
> Priority: Blocker
> Fix For: 2.5.2
>
> Attachments: AMBARI-21598.patch
>
>
> Steps to reproduce:-
> 1. Installed a IOP cluster ambari-version:- 2.2.0,BigInsights-4.2.0.0
> 2. Upgrade the ambari from 2.2.0 to 2.5.2.0-189(IOP Clusters)
> 3. Remove IOP Select.
> 4. Register HDP Stack to HDP-2.6.2.0-124
> 5. EU
> 6. Post EU
> Spark Thrift Server is failing after Upgrade:-
> Logs:-
> {code}
> 17/07/28 03:32:18 INFO SparkUI: Stopped Spark web UI at http://natr66-tbus-iop420tofnsec-r6-4.openstacklocal:4040
> 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Interrupting monitor thread
> 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Shutting down all executors
> 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Asking each executor to shut down
> 17/07/28 03:32:18 INFO SchedulerExtensionServices: Stopping SchedulerExtensionServices
> (serviceOption=None,
> services=List(),
> started=false)
> 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Stopped
> 17/07/28 03:32:18 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
> 17/07/28 03:32:18 INFO MemoryStore: MemoryStore cleared
> 17/07/28 03:32:18 INFO BlockManager: BlockManager stopped
> 17/07/28 03:32:18 INFO BlockManagerMaster: BlockManagerMaster stopped
> 17/07/28 03:32:18 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
> 17/07/28 03:32:18 INFO SparkContext: Successfully stopped SparkContext
> 17/07/28 03:32:18 ERROR Utils: Uncaught exception in thread pool-7-thread-1
> java.lang.NullPointerException
> at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$$anonfun$main$1.apply$mcV$sp(HiveThriftServer2.scala:123)
> at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267)
> at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239)
> at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
> at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
> at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1817)
> at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239)
> at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
> at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
> at scala.util.Try$.apply(Try.scala:161)
> at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239)
> at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> 17/07/28 03:32:18 INFO ShutdownHookManager: Shutdown hook called
> 17/07/28 03:32:18 INFO ShutdownHookManager: Deleting directory /tmp/spark-87997670-a290-4c52-a5f5-4ea0bbe87d4c
> 17/07/28 03:32:18 INFO ShutdownHookManager: Deleting directory /tmp/spark-f76d3d61-f7d5-4a0a-a50d-d3a1766f3f09
> 17/07/28 03:32:18 INFO RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon.
> 17/07/28 03:32:18 INFO RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports.
> {code}
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)