You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@jena.apache.org by Carmen Manzulli <ca...@gmail.com> on 2015/03/02 10:22:45 UTC

Re: error running rdfstats

ok...it was a problem due to the hadoop configuration files!thanks a
lot....now rdfstats starts! :D

2015-02-23 10:55 GMT+01:00 Rob Vesse <rv...@dotnetrdf.org>:

> If there job is stuck in a Pending state that is usually an issue with the
> configuration of your Hadoop cluster e.g. insufficient resources
>
> Those kinds of questions are best addressed to the Hadoop community or
> your Hadoop distribution vendor if you are using a non-Apache distribution
>
> Rob
>
> On 20/02/2015 14:07, "Carmen Manzulli" <ca...@gmail.com> wrote:
>
> >ok...now it seems to be able to start to execute the job but it rest in
> >pending status for a lot..what kind of problem could be ?
> >
> >2015-02-20 11:29 GMT+01:00 Carmen Manzulli <ca...@gmail.com>:
> >
> >> Thanks a lot....I was thinking about a Hadoop configuration
> >>problem..sorry
> >> for this question..
> >>
> >> 2015-02-19 19:01 GMT+01:00 Rob Vesse <rv...@dotnetrdf.org>:
> >>
> >>> --nodecount is not a valid argument so it is being treated as a file
> >>>name
> >>>
> >>> The correct argument is --node-count
> >>>
> >>> Rob
> >>>
> >>> On 19/02/2015 16:12, "Carmen Manzulli" <ca...@gmail.com>
> >>>wrote:
> >>>
> >>> >Hi,
> >>> >i don't know why but i'm getting this error:
> >>> >
> >>> >
> >>> >Input path does not exist:
> >>>hdfs://master:54310/user/sisinflab/--nodecount
> >>> >org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input
> >>>path
> >>> >does not exist: hdfs://master:54310/user/sisinflab/--nodecount
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.mapreduce.lib.input.FileInputFormat.singleThreadedLis
> >>>>tSt
> >>> >atus(FileInputFormat.java:321)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.mapreduce.lib.input.FileInputFormat.listStatus(FileIn
> >>>>put
> >>> >Format.java:264)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInp
> >>>>utF
> >>> >ormat.java:385)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.ja
> >>>>va:
> >>> >597)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:
> >>>>614
> >>> >)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter
> >>>>.ja
> >>> >va:492)
> >>> >    at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1296)
> >>> >    at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1293)
> >>> >    at java.security.AccessController.doPrivileged(Native Method)
> >>> >    at javax.security.auth.Subject.doAs(Subject.java:415)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformati
> >>>>on.
> >>> >java:1628)
> >>> >    at org.apache.hadoop.mapreduce.Job.submit(Job.java:1293)
> >>> >    at
> >>> org.apache.jena.hadoop.rdf.stats.RdfStats.runJob(RdfStats.java:297)
> >>> >    at
> >>>
> >>>
> >>>>org.apache.jena.hadoop.rdf.stats.RdfStats.computeStatistic(RdfStats.jav
> >>>>a:2
> >>> >78)
> >>> >    at
> >>>org.apache.jena.hadoop.rdf.stats.RdfStats.run(RdfStats.java:264)
> >>> >    at
> >>>org.apache.jena.hadoop.rdf.stats.RdfStats.run(RdfStats.java:180)
> >>> >    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >>> >    at
> >>>org.apache.jena.hadoop.rdf.stats.RdfStats.main(RdfStats.java:136)
> >>> >    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>> >    at
> >>>
> >>>
> >>>>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.ja
> >>>>va:
> >>> >57)
> >>> >    at
> >>>
> >>>
> >>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccesso
> >>>>rIm
> >>> >pl.java:43)
> >>> >    at java.lang.reflect.Method.invoke(Method.java:606)
> >>> >    at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
> >>> >    at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> >>> >
> >>> >
> >>> > when i try to run the command:
> >>> >
> >>> >hadoop jar
> >>>
> >>>
> >>>>/home/sisinflab/NetBeansProjects/prova/target/prova-1.0-SNAPSHOT-jar-wi
> >>>>th-
> >>> >dependencies.jar
> >>> >org.apache.jena.hadoop.rdf.stats.RdfStats --nodecount --data-types
> >>> >--namespaces --output hdfs:///prove/output --input-type triples
> >>> >hdfs:///input/persondata_en.ttl
> >>> >
> >>> >
> >>> >could anyone help me?
> >>> >thanks in advance
> >>>
> >>>
> >>>
> >>>
> >>>
> >>
>
>
>
>
>