You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Kal El <pi...@yahoo.com> on 2014/02/11 10:18:31 UTC
java.lang.ClassNotFoundException: org.apache.spark.streaming.StreamingContext
I am trying to run this code on spark 0.9.0 and I am getting this error: "java.lang.ClassNotFoundException: org.apache.spark.streaming.StreamingContext"
My CLASSPATH variable contains the following path"....spark-0.9.0/assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-hadoop1.0.4.jar"
what else do I need to add here ?
The code works fine in the spark-shell.
"import org.apache.spark.streaming.{Time, Seconds, StreamingContext}
import org.apache.spark.streaming.StreamingContext._
object HdfsWordCount {
def main(args: Array[String]) {
if (args.length < 2) {
System.err.println("Usage: HdfsWordCount <master> <directory>")
System.exit(1)
}
// StreamingExamples.setStreamingLogLevels()
// Create the context
// val ssc = new StreamingContext(args(0), "HdfsWordCount", Seconds(2), System.getenv("SPARK_HOME"), Seq(System.getenv("SPARK_HOME")))
val ssc = new StreamingContext(args(0), "HdfsWordCount", Seconds(2))
// Create the FileInputDStream on the directory and use the
// stream to count words in new files created
val lines = ssc.textFileStream(args(1))
val words = lines.flatMap(_.split(" "))
val wordCounts = words.map(x => (x, 1)).reduceByKey(_ + _)
wordCounts.print()
ssc.start()
ssc.awaitTermination()
}
}"
Thanks
Re: java.lang.ClassNotFoundException: org.apache.spark.streaming.StreamingContext
Posted by Amit Behera <am...@gmail.com>.
Hi Kal El,
to run you an application outside the spark-shell you need to write a *build
file* (example.sbt) parallel to your source code.
step1: *example.sbt*
name := "Simple Project"
version := "1.0"
scalaVersion := "2.9.3"
libraryDependencies += "org.apache.spark" %% "spark-streaming" %
"0.9.0-incubating"
resolvers ++= Seq( "Akka Repository" at
"http://repo.akka.io/releases/","Spray Repository" at
"http://repo.spray.cc/")
step2:
download sbt from
http://scalasbt.artifactoryonline.com/scalasbt/sbt-native-packages/org/scala-sbt/sbt/0.12.2/sbt.tgz
unzip it and set the path. (up to sbt/bin)
step3:
parallel to your .sbt file type command
$sbt package
$sbt run
[?]
On Tue, Feb 11, 2014 at 2:48 PM, Kal El <pi...@yahoo.com> wrote:
> I am trying to run this code on spark 0.9.0 and I am getting this error:
> "java.lang.ClassNotFoundException:
> org.apache.spark.streaming.StreamingContext"
>
> My CLASSPATH variable contains the following
> path"....spark-0.9.0/assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-hadoop1.0.4.jar"
> what else do I need to add here ?
>
> The code works fine in the spark-shell.
>
> "import org.apache.spark.streaming.{Time, Seconds, StreamingContext}
> import org.apache.spark.streaming.StreamingContext._
>
> object HdfsWordCount {
> def main(args: Array[String]) {
> if (args.length < 2) {
> System.err.println("Usage: HdfsWordCount <master> <directory>")
> System.exit(1)
> }
>
> // StreamingExamples.setStreamingLogLevels()
>
> // Create the context
> // val ssc = new StreamingContext(args(0), "HdfsWordCount", Seconds(2),
> System.getenv("SPARK_HOME"), Seq(System.getenv("SPARK_HOME")))
> val ssc = new StreamingContext(args(0), "HdfsWordCount", Seconds(2))
> // Create the FileInputDStream on the directory and use the
> // stream to count words in new files created
> val lines = ssc.textFileStream(args(1))
> val words = lines.flatMap(_.split(" "))
> val wordCounts = words.map(x => (x, 1)).reduceByKey(_ + _)
> wordCounts.print()
> ssc.start()
> ssc.awaitTermination()
> }
> }"
>
> Thanks
>
>
>