You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by ericacm <er...@gmail.com> on 2014/09/08 16:16:19 UTC

Cannot run SimpleApp as regular Java app

Dear all:

I am a brand new Spark user trying out the SimpleApp from the Quick Start
page.

Here is the code:

object SimpleApp {
  def main(args: Array[String]) {
    val logFile = "/dev/spark-1.0.2-bin-hadoop2/README.md" // Should be some
file on your system
    val conf = new SparkConf()
        .setAppName("Simple Application")
        .set("spark.executor.memory", "512m")
        .setMaster("spark://myhost.local:7077")
       
.setJars(Seq("/spark-experiments/target/spark-experiments-1.0-SNAPSHOT.jar"))
    val sc = new SparkContext(conf)
    try {
      val logData = sc.textFile(logFile, 2).cache()
      val numAs = logData.filter(line => line.contains("a")).count()
      val numBs = logData.filter(line => line.contains("b")).count()
      println("Lines with a: %s, Lines with b: %s".format(numAs, numBs))
    } finally {
      sc.stop()
    }
  }
}

I am using Spark 1.0.2 and Scala 2.10.4.  In spark-env.sh I have
SPARK_WORKER_MEMORY=2g.

I am trying to run this as a standalone Java app in my IDE.

Note that this code *does* work when I either
- Change the master to "local" (works running from IDE)
- Run it using spark-submit

The application/driver log is:

14/09/08 10:03:55 INFO spark.SecurityManager: Changing view acls to: eric
14/09/08 10:03:55 INFO spark.SecurityManager: SecurityManager:
authentication disabled; ui acls disabled; users with view permissions:
Set(eric)
14/09/08 10:03:56 INFO slf4j.Slf4jLogger: Slf4jLogger started
14/09/08 10:03:56 INFO Remoting: Starting remoting
14/09/08 10:03:56 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://spark@10.0.1.5:61645]
14/09/08 10:03:56 INFO Remoting: Remoting now listens on addresses:
[akka.tcp://spark@10.0.1.5:61645]
14/09/08 10:03:56 INFO spark.SparkEnv: Registering MapOutputTracker
14/09/08 10:03:56 INFO spark.SparkEnv: Registering BlockManagerMaster
14/09/08 10:03:56 INFO storage.DiskBlockManager: Created local directory at
/var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-local-20140908100356-2496
14/09/08 10:03:56 INFO storage.MemoryStore: MemoryStore started with
capacity 279.5 MB.
14/09/08 10:03:56 INFO network.ConnectionManager: Bound socket to port 61646
with id = ConnectionManagerId(10.0.1.5,61646)
14/09/08 10:03:56 INFO storage.BlockManagerMaster: Trying to register
BlockManager
14/09/08 10:03:56 INFO storage.BlockManagerInfo: Registering block manager
10.0.1.5:61646 with 279.5 MB RAM
14/09/08 10:03:56 INFO storage.BlockManagerMaster: Registered BlockManager
14/09/08 10:03:56 INFO spark.HttpServer: Starting HTTP Server
14/09/08 10:03:57 INFO server.Server: jetty-8.1.14.v20131031
14/09/08 10:03:57 INFO server.AbstractConnector: Started
SocketConnector@0.0.0.0:61647
14/09/08 10:03:57 INFO broadcast.HttpBroadcast: Broadcast server started at
http://10.0.1.5:61647
14/09/08 10:03:57 INFO spark.HttpFileServer: HTTP File server directory is
/var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-d5637279-5caa-4c14-a00f-650f1dd915bc
14/09/08 10:03:57 INFO spark.HttpServer: Starting HTTP Server
14/09/08 10:03:57 INFO server.Server: jetty-8.1.14.v20131031
14/09/08 10:03:57 INFO server.AbstractConnector: Started
SocketConnector@0.0.0.0:61648
14/09/08 10:03:57 INFO server.Server: jetty-8.1.14.v20131031
14/09/08 10:03:57 INFO server.AbstractConnector: Started
SelectChannelConnector@0.0.0.0:4040
14/09/08 10:03:57 INFO ui.SparkUI: Started SparkUI at http://10.0.1.5:4040
2014-09-08 10:03:57.567 java[58736:1703] Unable to load realm info from
SCDynamicStore
14/09/08 10:03:57 INFO spark.SparkContext: Added JAR
/spark-experiments/target/spark-experiments-1.0-SNAPSHOT.jar at
http://10.0.1.5:61648/jars/spark-experiments-1.0-SNAPSHOT.jar with timestamp
1410185037723
14/09/08 10:03:57 INFO client.AppClient$ClientActor: Connecting to master
spark://myhost.local:7077...
14/09/08 10:03:57 INFO storage.MemoryStore: ensureFreeSpace(32960) called
with curMem=0, maxMem=293063884
14/09/08 10:03:57 INFO storage.MemoryStore: Block broadcast_0 stored as
values to memory (estimated size 32.2 KB, free 279.5 MB)
14/09/08 10:03:58 INFO cluster.SparkDeploySchedulerBackend: Connected to
Spark cluster with app ID app-20140908100358-0002
14/09/08 10:03:58 INFO client.AppClient$ClientActor: Executor added:
app-20140908100358-0002/0 on worker-20140908100129-10.0.1.5-61526
(10.0.1.5:61526) with 8 cores
14/09/08 10:03:58 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20140908100358-0002/0 on hostPort 10.0.1.5:61526 with 8 cores, 512.0
MB RAM
14/09/08 10:03:58 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
14/09/08 10:03:58 WARN snappy.LoadSnappy: Snappy native library not loaded
14/09/08 10:03:58 INFO mapred.FileInputFormat: Total input paths to process
: 1
14/09/08 10:03:58 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/0 is now RUNNING
14/09/08 10:03:58 INFO spark.SparkContext: Starting job: count at
SimpleApp.scala:16
14/09/08 10:03:58 INFO scheduler.DAGScheduler: Got job 0 (count at
SimpleApp.scala:16) with 2 output partitions (allowLocal=false)
14/09/08 10:03:58 INFO scheduler.DAGScheduler: Final stage: Stage 0(count at
SimpleApp.scala:16)
14/09/08 10:03:58 INFO scheduler.DAGScheduler: Parents of final stage:
List()
14/09/08 10:03:58 INFO scheduler.DAGScheduler: Missing parents: List()
14/09/08 10:03:58 INFO scheduler.DAGScheduler: Submitting Stage 0
(FilteredRDD[2] at filter at SimpleApp.scala:16), which has no missing
parents
14/09/08 10:03:58 INFO scheduler.DAGScheduler: Submitting 2 missing tasks
from Stage 0 (FilteredRDD[2] at filter at SimpleApp.scala:16)
14/09/08 10:03:58 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 with
2 tasks
14/09/08 10:04:00 INFO cluster.SparkDeploySchedulerBackend: Registered
executor:
Actor[akka.tcp://sparkExecutor@10.0.1.5:61650/user/Executor#281350807] with
ID 0
14/09/08 10:04:01 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
0 on executor 0: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:01 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
1919 bytes in 2 ms
14/09/08 10:04:01 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
1 on executor 0: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:01 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
1919 bytes in 0 ms
14/09/08 10:04:01 INFO storage.BlockManagerInfo: Registering block manager
10.0.1.5:61655 with 294.4 MB RAM
14/09/08 10:04:02 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/0 is now EXITED (Command exited with code 52)
14/09/08 10:04:02 INFO cluster.SparkDeploySchedulerBackend: Executor 0
disconnected, so removing it
14/09/08 10:04:02 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20140908100358-0002/0 removed: Command exited with code 52
14/09/08 10:04:02 ERROR scheduler.TaskSchedulerImpl: Lost executor 0 on
10.0.1.5: remote Akka client disassociated
14/09/08 10:04:02 INFO scheduler.TaskSetManager: Re-queueing tasks for 0
from TaskSet 0.0
14/09/08 10:04:02 WARN scheduler.TaskSetManager: Lost TID 1 (task 0.0:1)
14/09/08 10:04:02 WARN scheduler.TaskSetManager: Lost TID 0 (task 0.0:0)
14/09/08 10:04:02 INFO client.AppClient$ClientActor: Executor added:
app-20140908100358-0002/1 on worker-20140908100129-10.0.1.5-61526
(10.0.1.5:61526) with 8 cores
14/09/08 10:04:02 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20140908100358-0002/1 on hostPort 10.0.1.5:61526 with 8 cores, 512.0
MB RAM
14/09/08 10:04:02 INFO scheduler.DAGScheduler: Executor lost: 0 (epoch 0)
14/09/08 10:04:02 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/1 is now RUNNING
14/09/08 10:04:02 INFO storage.BlockManagerMasterActor: Trying to remove
executor 0 from BlockManagerMaster.
14/09/08 10:04:02 INFO storage.BlockManagerMaster: Removed 0 successfully in
removeExecutor
14/09/08 10:04:05 INFO cluster.SparkDeploySchedulerBackend: Registered
executor:
Actor[akka.tcp://sparkExecutor@10.0.1.5:61668/user/Executor#162441524] with
ID 1
14/09/08 10:04:05 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
2 on executor 1: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:05 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
1919 bytes in 0 ms
14/09/08 10:04:05 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
3 on executor 1: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:05 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
1919 bytes in 0 ms
14/09/08 10:04:05 INFO storage.BlockManagerInfo: Registering block manager
10.0.1.5:61673 with 294.4 MB RAM
14/09/08 10:04:06 INFO cluster.SparkDeploySchedulerBackend: Executor 1
disconnected, so removing it
14/09/08 10:04:06 ERROR scheduler.TaskSchedulerImpl: Lost executor 1 on
10.0.1.5: remote Akka client disassociated
14/09/08 10:04:06 INFO scheduler.TaskSetManager: Re-queueing tasks for 1
from TaskSet 0.0
14/09/08 10:04:06 WARN scheduler.TaskSetManager: Lost TID 2 (task 0.0:0)
14/09/08 10:04:06 WARN scheduler.TaskSetManager: Lost TID 3 (task 0.0:1)
14/09/08 10:04:06 INFO scheduler.DAGScheduler: Executor lost: 1 (epoch 1)
14/09/08 10:04:06 INFO storage.BlockManagerMasterActor: Trying to remove
executor 1 from BlockManagerMaster.
14/09/08 10:04:06 INFO storage.BlockManagerMaster: Removed 1 successfully in
removeExecutor
14/09/08 10:04:06 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/1 is now EXITED (Command exited with code 52)
14/09/08 10:04:06 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20140908100358-0002/1 removed: Command exited with code 52
14/09/08 10:04:06 INFO client.AppClient$ClientActor: Executor added:
app-20140908100358-0002/2 on worker-20140908100129-10.0.1.5-61526
(10.0.1.5:61526) with 8 cores
14/09/08 10:04:06 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20140908100358-0002/2 on hostPort 10.0.1.5:61526 with 8 cores, 512.0
MB RAM
14/09/08 10:04:06 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/2 is now RUNNING
14/09/08 10:04:09 INFO cluster.SparkDeploySchedulerBackend: Registered
executor:
Actor[akka.tcp://sparkExecutor@10.0.1.5:61687/user/Executor#-1094963901]
with ID 2
14/09/08 10:04:09 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
4 on executor 2: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:09 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
1919 bytes in 0 ms
14/09/08 10:04:09 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
5 on executor 2: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:09 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
1919 bytes in 1 ms
14/09/08 10:04:10 INFO storage.BlockManagerInfo: Registering block manager
10.0.1.5:61692 with 294.4 MB RAM
14/09/08 10:04:11 INFO cluster.SparkDeploySchedulerBackend: Executor 2
disconnected, so removing it
14/09/08 10:04:11 ERROR scheduler.TaskSchedulerImpl: Lost executor 2 on
10.0.1.5: remote Akka client disassociated
14/09/08 10:04:11 INFO scheduler.TaskSetManager: Re-queueing tasks for 2
from TaskSet 0.0
14/09/08 10:04:11 WARN scheduler.TaskSetManager: Lost TID 5 (task 0.0:0)
14/09/08 10:04:11 WARN scheduler.TaskSetManager: Lost TID 4 (task 0.0:1)
14/09/08 10:04:11 INFO scheduler.DAGScheduler: Executor lost: 2 (epoch 2)
14/09/08 10:04:11 INFO storage.BlockManagerMasterActor: Trying to remove
executor 2 from BlockManagerMaster.
14/09/08 10:04:11 INFO storage.BlockManagerMaster: Removed 2 successfully in
removeExecutor
14/09/08 10:04:11 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/2 is now EXITED (Command exited with code 52)
14/09/08 10:04:11 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20140908100358-0002/2 removed: Command exited with code 52
14/09/08 10:04:11 INFO client.AppClient$ClientActor: Executor added:
app-20140908100358-0002/3 on worker-20140908100129-10.0.1.5-61526
(10.0.1.5:61526) with 8 cores
14/09/08 10:04:11 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20140908100358-0002/3 on hostPort 10.0.1.5:61526 with 8 cores, 512.0
MB RAM
14/09/08 10:04:11 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/3 is now RUNNING
14/09/08 10:04:14 INFO cluster.SparkDeploySchedulerBackend: Registered
executor:
Actor[akka.tcp://sparkExecutor@10.0.1.5:61705/user/Executor#-460934085] with
ID 3
14/09/08 10:04:14 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
6 on executor 3: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:14 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
1919 bytes in 0 ms
14/09/08 10:04:14 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
7 on executor 3: 10.0.1.5 (PROCESS_LOCAL)
14/09/08 10:04:14 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
1919 bytes in 1 ms
14/09/08 10:04:14 INFO storage.BlockManagerInfo: Registering block manager
10.0.1.5:61710 with 294.4 MB RAM
14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Executor 3
disconnected, so removing it
14/09/08 10:04:15 ERROR scheduler.TaskSchedulerImpl: Lost executor 3 on
10.0.1.5: remote Akka client disassociated
14/09/08 10:04:15 INFO scheduler.TaskSetManager: Re-queueing tasks for 3
from TaskSet 0.0
14/09/08 10:04:15 WARN scheduler.TaskSetManager: Lost TID 7 (task 0.0:0)
14/09/08 10:04:15 ERROR scheduler.TaskSetManager: Task 0.0:0 failed 4 times;
aborting job
14/09/08 10:04:15 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool 
14/09/08 10:04:15 INFO scheduler.TaskSchedulerImpl: Cancelling stage 0
14/09/08 10:04:15 INFO scheduler.DAGScheduler: Failed to run count at
SimpleApp.scala:16
14/09/08 10:04:15 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/3 is now EXITED (Command exited with code 52)
14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20140908100358-0002/3 removed: Command exited with code 52
14/09/08 10:04:15 INFO client.AppClient$ClientActor: Executor added:
app-20140908100358-0002/4 on worker-20140908100129-10.0.1.5-61526
(10.0.1.5:61526) with 8 cores
14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20140908100358-0002/4 on hostPort 10.0.1.5:61526 with 8 cores, 512.0
MB RAM
14/09/08 10:04:15 INFO scheduler.DAGScheduler: Executor lost: 3 (epoch 3)
14/09/08 10:04:15 INFO storage.BlockManagerMasterActor: Trying to remove
executor 3 from BlockManagerMaster.
14/09/08 10:04:15 INFO storage.BlockManagerMaster: Removed 3 successfully in
removeExecutor
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/metrics/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/stage/kill,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/static,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/executors/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/executors,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/environment/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/environment,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage/rdd/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage/rdd,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/storage,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/pool/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/pool,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/stage/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/stage,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages/json,null}
14/09/08 10:04:15 INFO handler.ContextHandler: stopped
o.e.j.s.ServletContextHandler{/stages,null}
14/09/08 10:04:15 INFO client.AppClient$ClientActor: Executor updated:
app-20140908100358-0002/4 is now RUNNING
14/09/08 10:04:15 INFO ui.SparkUI: Stopped Spark web UI at
http://10.0.1.5:4040
14/09/08 10:04:15 INFO scheduler.DAGScheduler: Stopping DAGScheduler
14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Shutting down
all executors
14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Asking each
executor to shut down
14/09/08 10:04:16 INFO spark.MapOutputTrackerMasterActor:
MapOutputTrackerActor stopped!
14/09/08 10:04:16 INFO network.ConnectionManager: Selector thread was
interrupted!
14/09/08 10:04:16 INFO network.ConnectionManager: ConnectionManager stopped
14/09/08 10:04:16 INFO storage.MemoryStore: MemoryStore cleared
14/09/08 10:04:16 INFO storage.BlockManager: BlockManager stopped
14/09/08 10:04:16 INFO storage.BlockManagerMasterActor: Stopping
BlockManagerMaster
14/09/08 10:04:17 INFO storage.BlockManagerMaster: BlockManagerMaster
stopped
14/09/08 10:04:17 INFO spark.SparkContext: Successfully stopped SparkContext
Exception in thread "main" org.apache.spark.SparkException: Job aborted due
to stage failure: Task 0.0:0 failed 4 times, most recent failure: TID 7 on
host 10.0.1.5 failed for unknown reason
Driver stacktrace:
	at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1049)
	at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1033)
	at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1031)
	at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
	at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1031)
	at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:635)
	at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:635)
	at scala.Option.foreach(Option.scala:236)
	at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:635)
	at
org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1234)
	at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
	at akka.actor.ActorCell.invoke(ActorCell.scala:456)
	at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
	at akka.dispatch.Mailbox.run(Mailbox.scala:219)
	at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
	at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
	at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
	at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
	at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
14/09/08 10:04:17 INFO remote.RemoteActorRefProvider$RemotingTerminator:
Shutting down remote daemon.

The stderr log from the executor is:

Spark Executor Command:
"/Library/Java/JavaVirtualMachines/jdk1.7.0_13.jdk/Contents/Home/bin/java"
"-cp"
"::/dev/spark-1.0.2-bin-hadoop2/conf:/dev/spark-1.0.2-bin-hadoop2/lib/spark-assembly-1.0.2-hadoop2.2.0.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-api-jdo-3.2.1.jar:/Users/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-core-3.2.2.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-rdbms-3.2.1.jar"
"-XX:MaxPermSize=128m" "-Xms512M" "-Xmx512M"
"org.apache.spark.executor.CoarseGrainedExecutorBackend"
"akka.tcp://spark@10.0.1.5:61645/user/CoarseGrainedScheduler" "1" "10.0.1.5"
"8" "akka.tcp://sparkWorker@10.0.1.5:61526/user/Worker"
"app-20140908100358-0002"
========================================

2014-09-08 10:04:04.202 java[58774:1703] Unable to load realm info from
SCDynamicStore
14/09/08 10:04:04 WARN NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
14/09/08 10:04:04 INFO SecurityManager: Changing view acls to: eric
14/09/08 10:04:04 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(eric)
14/09/08 10:04:05 INFO Slf4jLogger: Slf4jLogger started
14/09/08 10:04:05 INFO Remoting: Starting remoting
14/09/08 10:04:05 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://sparkExecutor@10.0.1.5:61668]
14/09/08 10:04:05 INFO Remoting: Remoting now listens on addresses:
[akka.tcp://sparkExecutor@10.0.1.5:61668]
14/09/08 10:04:05 INFO CoarseGrainedExecutorBackend: Connecting to driver:
akka.tcp://spark@10.0.1.5:61645/user/CoarseGrainedScheduler
14/09/08 10:04:05 INFO WorkerWatcher: Connecting to worker
akka.tcp://sparkWorker@10.0.1.5:61526/user/Worker
14/09/08 10:04:05 INFO WorkerWatcher: Successfully connected to
akka.tcp://sparkWorker@10.0.1.5:61526/user/Worker
14/09/08 10:04:05 INFO CoarseGrainedExecutorBackend: Successfully registered
with driver
14/09/08 10:04:05 INFO SecurityManager: Changing view acls to: eric
14/09/08 10:04:05 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(eric)
14/09/08 10:04:05 INFO Slf4jLogger: Slf4jLogger started
14/09/08 10:04:05 INFO Remoting: Starting remoting
14/09/08 10:04:05 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://spark@10.0.1.5:61671]
14/09/08 10:04:05 INFO Remoting: Remoting now listens on addresses:
[akka.tcp://spark@10.0.1.5:61671]
14/09/08 10:04:05 INFO SparkEnv: Connecting to MapOutputTracker:
akka.tcp://spark@10.0.1.5:61645/user/MapOutputTracker
14/09/08 10:04:05 INFO SparkEnv: Connecting to BlockManagerMaster:
akka.tcp://spark@10.0.1.5:61645/user/BlockManagerMaster
14/09/08 10:04:05 INFO DiskBlockManager: Created local directory at
/var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-local-20140908100405-509f
14/09/08 10:04:05 INFO MemoryStore: MemoryStore started with capacity 294.4
MB.
14/09/08 10:04:05 INFO ConnectionManager: Bound socket to port 61673 with id
= ConnectionManagerId(10.0.1.5,61673)
14/09/08 10:04:05 INFO BlockManagerMaster: Trying to register BlockManager
14/09/08 10:04:05 INFO BlockManagerMaster: Registered BlockManager
14/09/08 10:04:05 INFO HttpFileServer: HTTP File server directory is
/var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-3bb0c7fb-be2f-4efe-9dde-bcf8349a962a
14/09/08 10:04:05 INFO HttpServer: Starting HTTP Server
14/09/08 10:04:05 INFO CoarseGrainedExecutorBackend: Got assigned task 2
14/09/08 10:04:06 INFO CoarseGrainedExecutorBackend: Got assigned task 3
14/09/08 10:04:06 INFO Executor: Running task ID 2
14/09/08 10:04:06 INFO Executor: Running task ID 3
14/09/08 10:04:06 INFO Executor: Fetching
http://10.0.1.5:61648/jars/spark-experiments-1.0-SNAPSHOT.jar with timestamp
1410185037723
14/09/08 10:04:06 INFO Utils: Fetching
http://10.0.1.5:61648/jars/spark-experiments-1.0-SNAPSHOT.jar to
/var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/fetchFileTemp4194990807172315371.tmp
14/09/08 10:04:06 INFO Executor: Adding
file:/dev/spark-1.0.2-bin-hadoop2/work/app-20140908100358-0002/1/./spark-experiments-1.0-SNAPSHOT.jar
to class loader
14/09/08 10:04:06 INFO HttpBroadcast: Started reading broadcast variable 0
14/09/08 10:04:06 ERROR Executor: Exception in task ID 2
java.lang.OutOfMemoryError: Java heap space
	at
org.apache.hadoop.io.WritableUtils.readCompressedStringArray(WritableUtils.java:183)
	at org.apache.hadoop.conf.Configuration.readFields(Configuration.java:2378)
	at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:285)
	at org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:77)
	at
org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:42)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:601)
	at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
	at java.io.ObjectInputStream.readObject(ObjectInputStream.java:369)
	at
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:63)
	at org.apache.spark.broadcast.HttpBroadcast$.read(HttpBroadcast.scala:205)
	at
org.apache.spark.broadcast.HttpBroadcast.readObject(HttpBroadcast.scala:89)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:601)
	at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
14/09/08 10:04:06 INFO HttpBroadcast: Started reading broadcast variable 0
14/09/08 10:04:06 ERROR ExecutorUncaughtExceptionHandler: Uncaught exception
in thread Thread[Executor task launch worker-0,5,main]
java.lang.OutOfMemoryError: Java heap space
	at
org.apache.hadoop.io.WritableUtils.readCompressedStringArray(WritableUtils.java:183)
	at org.apache.hadoop.conf.Configuration.readFields(Configuration.java:2378)
	at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:285)
	at org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:77)
	at
org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:42)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:601)
	at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
	at java.io.ObjectInputStream.readObject(ObjectInputStream.java:369)
	at
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:63)
	at org.apache.spark.broadcast.HttpBroadcast$.read(HttpBroadcast.scala:205)
	at
org.apache.spark.broadcast.HttpBroadcast.readObject(HttpBroadcast.scala:89)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:601)
	at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
	at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)

It seems like a serialization problem because there is plenty of heap space
(and it works with spark-submit).

Thanks!



--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-run-SimpleApp-as-regular-Java-app-tp13695.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org


Re: Cannot run SimpleApp as regular Java app

Posted by ericacm <er...@gmail.com>.
It turns out that it was the Hadoop version that was the issue.

spark-1.0.2-hadoop1 and spark-1.1.0-hadoop1 both work.

spark.1.0.2-hadoop2, spark-1.1.0-hadoop2.4 and spark-1.1.0-hadoop2.4 do not
work.

It's strange because for this little test I am not even using HDFS at all.



-- Eric

On Thu, Sep 18, 2014 at 11:58 AM, ericacm [via Apache Spark User List] <
ml-node+s1001560n14570h61@n3.nabble.com> wrote:

> Upgrading from spark-1.0.2-hadoop2 to spark-1.1.0-hadoop1 fixed my
> problem.
>
> ------------------------------
>  If you reply to this email, your message will be added to the discussion
> below:
>
> http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-run-SimpleApp-as-regular-Java-app-tp13695p14570.html
>  To unsubscribe from Cannot run SimpleApp as regular Java app, click here
> <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=13695&code=ZXJpY2FjbUBnbWFpbC5jb218MTM2OTV8MTY0ODE0NDgzOQ==>
> .
> NAML
> <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>




--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-run-SimpleApp-as-regular-Java-app-tp13695p14685.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Re: Cannot run SimpleApp as regular Java app

Posted by ericacm <er...@gmail.com>.
Upgrading from spark-1.0.2-hadoop2 to spark-1.1.0-hadoop1 fixed my problem. 



--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-run-SimpleApp-as-regular-Java-app-tp13695p14570.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org


Re: Cannot run SimpleApp as regular Java app

Posted by ericacm <er...@gmail.com>.
Hi Yana - 

I added the following to spark-class:

echo RUNNER: $RUNNER
echo CLASSPATH: $CLASSPATH
echo JAVA_OPTS: $JAVA_OPTS
echo '$@': $@

Here's the output:

$ ./spark-submit --class experiments.SimpleApp --master
spark://myhost.local:7077
/IdeaProjects/spark-experiments/target/spark-experiments-1.0-SNAPSHOT.jar

Spark assembly has been built with Hive, including Datanucleus jars on
classpath

RUNNER:
/Library/Java/JavaVirtualMachines/jdk1.7.0_13.jdk/Contents/Home/bin/java

CLASSPATH:
::/dev/spark-1.0.2-bin-hadoop2/conf:/dev/spark-1.0.2-bin-hadoop2/lib/spark-assembly-1.0.2-hadoop2.2.0.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-api-jdo-3.2.1.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-core-3.2.2.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-rdbms-3.2.1.jar

JAVA_OPTS: -XX:MaxPermSize=128m -Djava.library.path= -Xms512m -Xmx512m

$@: org.apache.spark.deploy.SparkSubmit --class experiments.SimpleApp
--master spark://myhost.local:7077
/IdeaProjects/spark-experiments/target/spark-experiments-1.0-SNAPSHOT.jar

The differences I can see in the code that runs via my standalone Java app:
- Does not have -Djava.library.path=  (should not make a difference)
- Main class is org.apache.spark.executor.CoarseGrainedExecutorBackend
instead of org.apache.spark.deploy.SparkSubmit (should not make a
difference)
- My jar's classes are directly available when running via spark-submit (it
runs the Jar so it they will be in the main classloader) but they are only
available via conf.setJars() in the standalone Java app.  But they should be
available indirectly in the classloader that is created in the executor:

14/09/08 10:04:06 INFO Executor: Adding
file:/dev/spark-1.0.2-bin-hadoop2/work/app-20140908100358-0002/1/./spark-experiments-1.0-SNAPSHOT.jar
to class loader

I've been assuming that my conf.setJars() is the proper way to provide my
code to Spark.  

Thanks!




--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-run-SimpleApp-as-regular-Java-app-tp13695p13842.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org


Re: Cannot run SimpleApp as regular Java app

Posted by Yana Kadiyska <ya...@gmail.com>.
spark-submit is a script which calls spark-class script. Can you output the
command that spark-class runs (say, by putting set -x before the very last
line?). You should see the java command that is being run. The scripts do
some parameter setting so it's possible you're missing something. It seems
to me you think your worker memory is 2G but the executor is clearly
launched with -Xms512M" "-Xmx512M"...so that's all you'd get.

On Mon, Sep 8, 2014 at 10:16 AM, ericacm <er...@gmail.com> wrote:

> Dear all:
>
> I am a brand new Spark user trying out the SimpleApp from the Quick Start
> page.
>
> Here is the code:
>
> object SimpleApp {
>   def main(args: Array[String]) {
>     val logFile = "/dev/spark-1.0.2-bin-hadoop2/README.md" // Should be
> some
> file on your system
>     val conf = new SparkConf()
>         .setAppName("Simple Application")
>         .set("spark.executor.memory", "512m")
>         .setMaster("spark://myhost.local:7077")
>
>
> .setJars(Seq("/spark-experiments/target/spark-experiments-1.0-SNAPSHOT.jar"))
>     val sc = new SparkContext(conf)
>     try {
>       val logData = sc.textFile(logFile, 2).cache()
>       val numAs = logData.filter(line => line.contains("a")).count()
>       val numBs = logData.filter(line => line.contains("b")).count()
>       println("Lines with a: %s, Lines with b: %s".format(numAs, numBs))
>     } finally {
>       sc.stop()
>     }
>   }
> }
>
> I am using Spark 1.0.2 and Scala 2.10.4.  In spark-env.sh I have
> SPARK_WORKER_MEMORY=2g.
>
> I am trying to run this as a standalone Java app in my IDE.
>
> Note that this code *does* work when I either
> - Change the master to "local" (works running from IDE)
> - Run it using spark-submit
>
> The application/driver log is:
>
> 14/09/08 10:03:55 INFO spark.SecurityManager: Changing view acls to: eric
> 14/09/08 10:03:55 INFO spark.SecurityManager: SecurityManager:
> authentication disabled; ui acls disabled; users with view permissions:
> Set(eric)
> 14/09/08 10:03:56 INFO slf4j.Slf4jLogger: Slf4jLogger started
> 14/09/08 10:03:56 INFO Remoting: Starting remoting
> 14/09/08 10:03:56 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://spark@10.0.1.5:61645]
> 14/09/08 10:03:56 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://spark@10.0.1.5:61645]
> 14/09/08 10:03:56 INFO spark.SparkEnv: Registering MapOutputTracker
> 14/09/08 10:03:56 INFO spark.SparkEnv: Registering BlockManagerMaster
> 14/09/08 10:03:56 INFO storage.DiskBlockManager: Created local directory at
>
> /var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-local-20140908100356-2496
> 14/09/08 10:03:56 INFO storage.MemoryStore: MemoryStore started with
> capacity 279.5 MB.
> 14/09/08 10:03:56 INFO network.ConnectionManager: Bound socket to port
> 61646
> with id = ConnectionManagerId(10.0.1.5,61646)
> 14/09/08 10:03:56 INFO storage.BlockManagerMaster: Trying to register
> BlockManager
> 14/09/08 10:03:56 INFO storage.BlockManagerInfo: Registering block manager
> 10.0.1.5:61646 with 279.5 MB RAM
> 14/09/08 10:03:56 INFO storage.BlockManagerMaster: Registered BlockManager
> 14/09/08 10:03:56 INFO spark.HttpServer: Starting HTTP Server
> 14/09/08 10:03:57 INFO server.Server: jetty-8.1.14.v20131031
> 14/09/08 10:03:57 INFO server.AbstractConnector: Started
> SocketConnector@0.0.0.0:61647
> 14/09/08 10:03:57 INFO broadcast.HttpBroadcast: Broadcast server started at
> http://10.0.1.5:61647
> 14/09/08 10:03:57 INFO spark.HttpFileServer: HTTP File server directory is
>
> /var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-d5637279-5caa-4c14-a00f-650f1dd915bc
> 14/09/08 10:03:57 INFO spark.HttpServer: Starting HTTP Server
> 14/09/08 10:03:57 INFO server.Server: jetty-8.1.14.v20131031
> 14/09/08 10:03:57 INFO server.AbstractConnector: Started
> SocketConnector@0.0.0.0:61648
> 14/09/08 10:03:57 INFO server.Server: jetty-8.1.14.v20131031
> 14/09/08 10:03:57 INFO server.AbstractConnector: Started
> SelectChannelConnector@0.0.0.0:4040
> 14/09/08 10:03:57 INFO ui.SparkUI: Started SparkUI at http://10.0.1.5:4040
> 2014-09-08 10:03:57.567 java[58736:1703] Unable to load realm info from
> SCDynamicStore
> 14/09/08 10:03:57 INFO spark.SparkContext: Added JAR
> /spark-experiments/target/spark-experiments-1.0-SNAPSHOT.jar at
> http://10.0.1.5:61648/jars/spark-experiments-1.0-SNAPSHOT.jar with
> timestamp
> 1410185037723
> 14/09/08 10:03:57 INFO client.AppClient$ClientActor: Connecting to master
> spark://myhost.local:7077...
> 14/09/08 10:03:57 INFO storage.MemoryStore: ensureFreeSpace(32960) called
> with curMem=0, maxMem=293063884
> 14/09/08 10:03:57 INFO storage.MemoryStore: Block broadcast_0 stored as
> values to memory (estimated size 32.2 KB, free 279.5 MB)
> 14/09/08 10:03:58 INFO cluster.SparkDeploySchedulerBackend: Connected to
> Spark cluster with app ID app-20140908100358-0002
> 14/09/08 10:03:58 INFO client.AppClient$ClientActor: Executor added:
> app-20140908100358-0002/0 on worker-20140908100129-10.0.1.5-61526
> (10.0.1.5:61526) with 8 cores
> 14/09/08 10:03:58 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20140908100358-0002/0 on hostPort 10.0.1.5:61526 with 8 cores,
> 512.0
> MB RAM
> 14/09/08 10:03:58 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 14/09/08 10:03:58 WARN snappy.LoadSnappy: Snappy native library not loaded
> 14/09/08 10:03:58 INFO mapred.FileInputFormat: Total input paths to process
> : 1
> 14/09/08 10:03:58 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/0 is now RUNNING
> 14/09/08 10:03:58 INFO spark.SparkContext: Starting job: count at
> SimpleApp.scala:16
> 14/09/08 10:03:58 INFO scheduler.DAGScheduler: Got job 0 (count at
> SimpleApp.scala:16) with 2 output partitions (allowLocal=false)
> 14/09/08 10:03:58 INFO scheduler.DAGScheduler: Final stage: Stage 0(count
> at
> SimpleApp.scala:16)
> 14/09/08 10:03:58 INFO scheduler.DAGScheduler: Parents of final stage:
> List()
> 14/09/08 10:03:58 INFO scheduler.DAGScheduler: Missing parents: List()
> 14/09/08 10:03:58 INFO scheduler.DAGScheduler: Submitting Stage 0
> (FilteredRDD[2] at filter at SimpleApp.scala:16), which has no missing
> parents
> 14/09/08 10:03:58 INFO scheduler.DAGScheduler: Submitting 2 missing tasks
> from Stage 0 (FilteredRDD[2] at filter at SimpleApp.scala:16)
> 14/09/08 10:03:58 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0
> with
> 2 tasks
> 14/09/08 10:04:00 INFO cluster.SparkDeploySchedulerBackend: Registered
> executor:
> Actor[akka.tcp://sparkExecutor@10.0.1.5:61650/user/Executor#281350807]
> with
> ID 0
> 14/09/08 10:04:01 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
> 0 on executor 0: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:01 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
> 1919 bytes in 2 ms
> 14/09/08 10:04:01 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
> 1 on executor 0: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:01 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
> 1919 bytes in 0 ms
> 14/09/08 10:04:01 INFO storage.BlockManagerInfo: Registering block manager
> 10.0.1.5:61655 with 294.4 MB RAM
> 14/09/08 10:04:02 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/0 is now EXITED (Command exited with code 52)
> 14/09/08 10:04:02 INFO cluster.SparkDeploySchedulerBackend: Executor 0
> disconnected, so removing it
> 14/09/08 10:04:02 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20140908100358-0002/0 removed: Command exited with code 52
> 14/09/08 10:04:02 ERROR scheduler.TaskSchedulerImpl: Lost executor 0 on
> 10.0.1.5: remote Akka client disassociated
> 14/09/08 10:04:02 INFO scheduler.TaskSetManager: Re-queueing tasks for 0
> from TaskSet 0.0
> 14/09/08 10:04:02 WARN scheduler.TaskSetManager: Lost TID 1 (task 0.0:1)
> 14/09/08 10:04:02 WARN scheduler.TaskSetManager: Lost TID 0 (task 0.0:0)
> 14/09/08 10:04:02 INFO client.AppClient$ClientActor: Executor added:
> app-20140908100358-0002/1 on worker-20140908100129-10.0.1.5-61526
> (10.0.1.5:61526) with 8 cores
> 14/09/08 10:04:02 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20140908100358-0002/1 on hostPort 10.0.1.5:61526 with 8 cores,
> 512.0
> MB RAM
> 14/09/08 10:04:02 INFO scheduler.DAGScheduler: Executor lost: 0 (epoch 0)
> 14/09/08 10:04:02 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/1 is now RUNNING
> 14/09/08 10:04:02 INFO storage.BlockManagerMasterActor: Trying to remove
> executor 0 from BlockManagerMaster.
> 14/09/08 10:04:02 INFO storage.BlockManagerMaster: Removed 0 successfully
> in
> removeExecutor
> 14/09/08 10:04:05 INFO cluster.SparkDeploySchedulerBackend: Registered
> executor:
> Actor[akka.tcp://sparkExecutor@10.0.1.5:61668/user/Executor#162441524]
> with
> ID 1
> 14/09/08 10:04:05 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
> 2 on executor 1: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:05 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
> 1919 bytes in 0 ms
> 14/09/08 10:04:05 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
> 3 on executor 1: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:05 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
> 1919 bytes in 0 ms
> 14/09/08 10:04:05 INFO storage.BlockManagerInfo: Registering block manager
> 10.0.1.5:61673 with 294.4 MB RAM
> 14/09/08 10:04:06 INFO cluster.SparkDeploySchedulerBackend: Executor 1
> disconnected, so removing it
> 14/09/08 10:04:06 ERROR scheduler.TaskSchedulerImpl: Lost executor 1 on
> 10.0.1.5: remote Akka client disassociated
> 14/09/08 10:04:06 INFO scheduler.TaskSetManager: Re-queueing tasks for 1
> from TaskSet 0.0
> 14/09/08 10:04:06 WARN scheduler.TaskSetManager: Lost TID 2 (task 0.0:0)
> 14/09/08 10:04:06 WARN scheduler.TaskSetManager: Lost TID 3 (task 0.0:1)
> 14/09/08 10:04:06 INFO scheduler.DAGScheduler: Executor lost: 1 (epoch 1)
> 14/09/08 10:04:06 INFO storage.BlockManagerMasterActor: Trying to remove
> executor 1 from BlockManagerMaster.
> 14/09/08 10:04:06 INFO storage.BlockManagerMaster: Removed 1 successfully
> in
> removeExecutor
> 14/09/08 10:04:06 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/1 is now EXITED (Command exited with code 52)
> 14/09/08 10:04:06 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20140908100358-0002/1 removed: Command exited with code 52
> 14/09/08 10:04:06 INFO client.AppClient$ClientActor: Executor added:
> app-20140908100358-0002/2 on worker-20140908100129-10.0.1.5-61526
> (10.0.1.5:61526) with 8 cores
> 14/09/08 10:04:06 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20140908100358-0002/2 on hostPort 10.0.1.5:61526 with 8 cores,
> 512.0
> MB RAM
> 14/09/08 10:04:06 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/2 is now RUNNING
> 14/09/08 10:04:09 INFO cluster.SparkDeploySchedulerBackend: Registered
> executor:
> Actor[akka.tcp://sparkExecutor@10.0.1.5:61687/user/Executor#-1094963901]
> with ID 2
> 14/09/08 10:04:09 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
> 4 on executor 2: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:09 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
> 1919 bytes in 0 ms
> 14/09/08 10:04:09 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
> 5 on executor 2: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:09 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
> 1919 bytes in 1 ms
> 14/09/08 10:04:10 INFO storage.BlockManagerInfo: Registering block manager
> 10.0.1.5:61692 with 294.4 MB RAM
> 14/09/08 10:04:11 INFO cluster.SparkDeploySchedulerBackend: Executor 2
> disconnected, so removing it
> 14/09/08 10:04:11 ERROR scheduler.TaskSchedulerImpl: Lost executor 2 on
> 10.0.1.5: remote Akka client disassociated
> 14/09/08 10:04:11 INFO scheduler.TaskSetManager: Re-queueing tasks for 2
> from TaskSet 0.0
> 14/09/08 10:04:11 WARN scheduler.TaskSetManager: Lost TID 5 (task 0.0:0)
> 14/09/08 10:04:11 WARN scheduler.TaskSetManager: Lost TID 4 (task 0.0:1)
> 14/09/08 10:04:11 INFO scheduler.DAGScheduler: Executor lost: 2 (epoch 2)
> 14/09/08 10:04:11 INFO storage.BlockManagerMasterActor: Trying to remove
> executor 2 from BlockManagerMaster.
> 14/09/08 10:04:11 INFO storage.BlockManagerMaster: Removed 2 successfully
> in
> removeExecutor
> 14/09/08 10:04:11 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/2 is now EXITED (Command exited with code 52)
> 14/09/08 10:04:11 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20140908100358-0002/2 removed: Command exited with code 52
> 14/09/08 10:04:11 INFO client.AppClient$ClientActor: Executor added:
> app-20140908100358-0002/3 on worker-20140908100129-10.0.1.5-61526
> (10.0.1.5:61526) with 8 cores
> 14/09/08 10:04:11 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20140908100358-0002/3 on hostPort 10.0.1.5:61526 with 8 cores,
> 512.0
> MB RAM
> 14/09/08 10:04:11 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/3 is now RUNNING
> 14/09/08 10:04:14 INFO cluster.SparkDeploySchedulerBackend: Registered
> executor:
> Actor[akka.tcp://sparkExecutor@10.0.1.5:61705/user/Executor#-460934085]
> with
> ID 3
> 14/09/08 10:04:14 INFO scheduler.TaskSetManager: Starting task 0.0:1 as TID
> 6 on executor 3: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:14 INFO scheduler.TaskSetManager: Serialized task 0.0:1 as
> 1919 bytes in 0 ms
> 14/09/08 10:04:14 INFO scheduler.TaskSetManager: Starting task 0.0:0 as TID
> 7 on executor 3: 10.0.1.5 (PROCESS_LOCAL)
> 14/09/08 10:04:14 INFO scheduler.TaskSetManager: Serialized task 0.0:0 as
> 1919 bytes in 1 ms
> 14/09/08 10:04:14 INFO storage.BlockManagerInfo: Registering block manager
> 10.0.1.5:61710 with 294.4 MB RAM
> 14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Executor 3
> disconnected, so removing it
> 14/09/08 10:04:15 ERROR scheduler.TaskSchedulerImpl: Lost executor 3 on
> 10.0.1.5: remote Akka client disassociated
> 14/09/08 10:04:15 INFO scheduler.TaskSetManager: Re-queueing tasks for 3
> from TaskSet 0.0
> 14/09/08 10:04:15 WARN scheduler.TaskSetManager: Lost TID 7 (task 0.0:0)
> 14/09/08 10:04:15 ERROR scheduler.TaskSetManager: Task 0.0:0 failed 4
> times;
> aborting job
> 14/09/08 10:04:15 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
> whose tasks have all completed, from pool
> 14/09/08 10:04:15 INFO scheduler.TaskSchedulerImpl: Cancelling stage 0
> 14/09/08 10:04:15 INFO scheduler.DAGScheduler: Failed to run count at
> SimpleApp.scala:16
> 14/09/08 10:04:15 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/3 is now EXITED (Command exited with code 52)
> 14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20140908100358-0002/3 removed: Command exited with code 52
> 14/09/08 10:04:15 INFO client.AppClient$ClientActor: Executor added:
> app-20140908100358-0002/4 on worker-20140908100129-10.0.1.5-61526
> (10.0.1.5:61526) with 8 cores
> 14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20140908100358-0002/4 on hostPort 10.0.1.5:61526 with 8 cores,
> 512.0
> MB RAM
> 14/09/08 10:04:15 INFO scheduler.DAGScheduler: Executor lost: 3 (epoch 3)
> 14/09/08 10:04:15 INFO storage.BlockManagerMasterActor: Trying to remove
> executor 3 from BlockManagerMaster.
> 14/09/08 10:04:15 INFO storage.BlockManagerMaster: Removed 3 successfully
> in
> removeExecutor
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/metrics/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages/stage/kill,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/static,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/executors/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/executors,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/environment/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/environment,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/storage/rdd/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/storage/rdd,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/storage/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/storage,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages/pool/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages/pool,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages/stage/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages/stage,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages/json,null}
> 14/09/08 10:04:15 INFO handler.ContextHandler: stopped
> o.e.j.s.ServletContextHandler{/stages,null}
> 14/09/08 10:04:15 INFO client.AppClient$ClientActor: Executor updated:
> app-20140908100358-0002/4 is now RUNNING
> 14/09/08 10:04:15 INFO ui.SparkUI: Stopped Spark web UI at
> http://10.0.1.5:4040
> 14/09/08 10:04:15 INFO scheduler.DAGScheduler: Stopping DAGScheduler
> 14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Shutting down
> all executors
> 14/09/08 10:04:15 INFO cluster.SparkDeploySchedulerBackend: Asking each
> executor to shut down
> 14/09/08 10:04:16 INFO spark.MapOutputTrackerMasterActor:
> MapOutputTrackerActor stopped!
> 14/09/08 10:04:16 INFO network.ConnectionManager: Selector thread was
> interrupted!
> 14/09/08 10:04:16 INFO network.ConnectionManager: ConnectionManager stopped
> 14/09/08 10:04:16 INFO storage.MemoryStore: MemoryStore cleared
> 14/09/08 10:04:16 INFO storage.BlockManager: BlockManager stopped
> 14/09/08 10:04:16 INFO storage.BlockManagerMasterActor: Stopping
> BlockManagerMaster
> 14/09/08 10:04:17 INFO storage.BlockManagerMaster: BlockManagerMaster
> stopped
> 14/09/08 10:04:17 INFO spark.SparkContext: Successfully stopped
> SparkContext
> Exception in thread "main" org.apache.spark.SparkException: Job aborted due
> to stage failure: Task 0.0:0 failed 4 times, most recent failure: TID 7 on
> host 10.0.1.5 failed for unknown reason
> Driver stacktrace:
>         at
> org.apache.spark.scheduler.DAGScheduler.org
> $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1049)
>         at
>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1033)
>         at
>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1031)
>         at
>
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>         at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
>         at
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1031)
>         at
>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:635)
>         at
>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:635)
>         at scala.Option.foreach(Option.scala:236)
>         at
>
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:635)
>         at
>
> org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1234)
>         at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
>         at akka.actor.ActorCell.invoke(ActorCell.scala:456)
>         at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
>         at akka.dispatch.Mailbox.run(Mailbox.scala:219)
>         at
>
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
>         at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>         at
>
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>         at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>         at
>
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> 14/09/08 10:04:17 INFO remote.RemoteActorRefProvider$RemotingTerminator:
> Shutting down remote daemon.
>
> The stderr log from the executor is:
>
> Spark Executor Command:
> "/Library/Java/JavaVirtualMachines/jdk1.7.0_13.jdk/Contents/Home/bin/java"
> "-cp"
>
> "::/dev/spark-1.0.2-bin-hadoop2/conf:/dev/spark-1.0.2-bin-hadoop2/lib/spark-assembly-1.0.2-hadoop2.2.0.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-api-jdo-3.2.1.jar:/Users/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-core-3.2.2.jar:/dev/spark-1.0.2-bin-hadoop2/lib/datanucleus-rdbms-3.2.1.jar"
> "-XX:MaxPermSize=128m" "-Xms512M" "-Xmx512M"
> "org.apache.spark.executor.CoarseGrainedExecutorBackend"
> "akka.tcp://spark@10.0.1.5:61645/user/CoarseGrainedScheduler" "1"
> "10.0.1.5"
> "8" "akka.tcp://sparkWorker@10.0.1.5:61526/user/Worker"
> "app-20140908100358-0002"
> ========================================
>
> 2014-09-08 10:04:04.202 java[58774:1703] Unable to load realm info from
> SCDynamicStore
> 14/09/08 10:04:04 WARN NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 14/09/08 10:04:04 INFO SecurityManager: Changing view acls to: eric
> 14/09/08 10:04:04 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions: Set(eric)
> 14/09/08 10:04:05 INFO Slf4jLogger: Slf4jLogger started
> 14/09/08 10:04:05 INFO Remoting: Starting remoting
> 14/09/08 10:04:05 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://sparkExecutor@10.0.1.5:61668]
> 14/09/08 10:04:05 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://sparkExecutor@10.0.1.5:61668]
> 14/09/08 10:04:05 INFO CoarseGrainedExecutorBackend: Connecting to driver:
> akka.tcp://spark@10.0.1.5:61645/user/CoarseGrainedScheduler
> 14/09/08 10:04:05 INFO WorkerWatcher: Connecting to worker
> akka.tcp://sparkWorker@10.0.1.5:61526/user/Worker
> 14/09/08 10:04:05 INFO WorkerWatcher: Successfully connected to
> akka.tcp://sparkWorker@10.0.1.5:61526/user/Worker
> 14/09/08 10:04:05 INFO CoarseGrainedExecutorBackend: Successfully
> registered
> with driver
> 14/09/08 10:04:05 INFO SecurityManager: Changing view acls to: eric
> 14/09/08 10:04:05 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions: Set(eric)
> 14/09/08 10:04:05 INFO Slf4jLogger: Slf4jLogger started
> 14/09/08 10:04:05 INFO Remoting: Starting remoting
> 14/09/08 10:04:05 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://spark@10.0.1.5:61671]
> 14/09/08 10:04:05 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://spark@10.0.1.5:61671]
> 14/09/08 10:04:05 INFO SparkEnv: Connecting to MapOutputTracker:
> akka.tcp://spark@10.0.1.5:61645/user/MapOutputTracker
> 14/09/08 10:04:05 INFO SparkEnv: Connecting to BlockManagerMaster:
> akka.tcp://spark@10.0.1.5:61645/user/BlockManagerMaster
> 14/09/08 10:04:05 INFO DiskBlockManager: Created local directory at
>
> /var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-local-20140908100405-509f
> 14/09/08 10:04:05 INFO MemoryStore: MemoryStore started with capacity 294.4
> MB.
> 14/09/08 10:04:05 INFO ConnectionManager: Bound socket to port 61673 with
> id
> = ConnectionManagerId(10.0.1.5,61673)
> 14/09/08 10:04:05 INFO BlockManagerMaster: Trying to register BlockManager
> 14/09/08 10:04:05 INFO BlockManagerMaster: Registered BlockManager
> 14/09/08 10:04:05 INFO HttpFileServer: HTTP File server directory is
>
> /var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/spark-3bb0c7fb-be2f-4efe-9dde-bcf8349a962a
> 14/09/08 10:04:05 INFO HttpServer: Starting HTTP Server
> 14/09/08 10:04:05 INFO CoarseGrainedExecutorBackend: Got assigned task 2
> 14/09/08 10:04:06 INFO CoarseGrainedExecutorBackend: Got assigned task 3
> 14/09/08 10:04:06 INFO Executor: Running task ID 2
> 14/09/08 10:04:06 INFO Executor: Running task ID 3
> 14/09/08 10:04:06 INFO Executor: Fetching
> http://10.0.1.5:61648/jars/spark-experiments-1.0-SNAPSHOT.jar with
> timestamp
> 1410185037723
> 14/09/08 10:04:06 INFO Utils: Fetching
> http://10.0.1.5:61648/jars/spark-experiments-1.0-SNAPSHOT.jar to
>
> /var/folders/j1/5rzyf1x97q9_7gj3mdc79t3c0000gn/T/fetchFileTemp4194990807172315371.tmp
> 14/09/08 10:04:06 INFO Executor: Adding
>
> file:/dev/spark-1.0.2-bin-hadoop2/work/app-20140908100358-0002/1/./spark-experiments-1.0-SNAPSHOT.jar
> to class loader
> 14/09/08 10:04:06 INFO HttpBroadcast: Started reading broadcast variable 0
> 14/09/08 10:04:06 ERROR Executor: Exception in task ID 2
> java.lang.OutOfMemoryError: Java heap space
>         at
>
> org.apache.hadoop.io.WritableUtils.readCompressedStringArray(WritableUtils.java:183)
>         at
> org.apache.hadoop.conf.Configuration.readFields(Configuration.java:2378)
>         at
> org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:285)
>         at
> org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:77)
>         at
>
> org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:42)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:601)
>         at
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
>         at java.io.ObjectInputStream.readObject(ObjectInputStream.java:369)
>         at
>
> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:63)
>         at
> org.apache.spark.broadcast.HttpBroadcast$.read(HttpBroadcast.scala:205)
>         at
> org.apache.spark.broadcast.HttpBroadcast.readObject(HttpBroadcast.scala:89)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:601)
>         at
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
> 14/09/08 10:04:06 INFO HttpBroadcast: Started reading broadcast variable 0
> 14/09/08 10:04:06 ERROR ExecutorUncaughtExceptionHandler: Uncaught
> exception
> in thread Thread[Executor task launch worker-0,5,main]
> java.lang.OutOfMemoryError: Java heap space
>         at
>
> org.apache.hadoop.io.WritableUtils.readCompressedStringArray(WritableUtils.java:183)
>         at
> org.apache.hadoop.conf.Configuration.readFields(Configuration.java:2378)
>         at
> org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:285)
>         at
> org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:77)
>         at
>
> org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:42)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:601)
>         at
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
>         at java.io.ObjectInputStream.readObject(ObjectInputStream.java:369)
>         at
>
> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:63)
>         at
> org.apache.spark.broadcast.HttpBroadcast$.read(HttpBroadcast.scala:205)
>         at
> org.apache.spark.broadcast.HttpBroadcast.readObject(HttpBroadcast.scala:89)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:601)
>         at
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1004)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1872)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1347)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1970)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1894)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1777)
>
> It seems like a serialization problem because there is plenty of heap space
> (and it works with spark-submit).
>
> Thanks!
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-run-SimpleApp-as-regular-Java-app-tp13695.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
> For additional commands, e-mail: user-help@spark.apache.org
>
>