You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by Robert Irribarren <ro...@algorithms.io> on 2012/08/26 06:25:53 UTC

Nutch 2.0 error

org.apache.solr.common.SolrException: Server Error

Server Error

request: http://127.0.0.1:8983/solr/sites/update?wt=javabin&version=2
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:430)
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244)
	at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
	at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:49)
	at org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:92)
	at org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:52)
	at org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.close(MapTask.java:651)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:766)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
	at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
	at org.apache.hadoop.mapred.Child.main(Child.java:249)

org.apache.solr.common.SolrException: Server Error

Server Error

request: http://127.0.0.1:8983/solr/sites/update?wt=javabin&version=2
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:430)
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244)
	at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
	at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:49)
	at org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:92)
	at org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:52)
	at org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.close(MapTask.java:651)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:766)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
	at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
	at org.apache.hadoop.mapred.Child.main(Child.java:249)

org.apache.solr.common.SolrException: Server Error

Server Error

request: http://127.0.0.1:8983/solr/sites/update?wt=javabin&version=2
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:430)
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244)
	at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
	at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:49)
	at org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:92)
	at org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:52)
	at org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.close(MapTask.java:651)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:766)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
	at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
	at org.apache.hadoop.mapred.Child.main(Child.java:249)

org.apache.solr.common.SolrException: Server Error

Server Error

request: http://127.0.0.1:8983/solr/sites/update?wt=javabin&version=2
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:430)
	at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244)
	at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
	at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:49)
	at org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:92)
	at org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:52)
	at org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.close(MapTask.java:651)
	at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:766)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
	at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
	at org.apache.hadoop.mapred.Child.main(Child.java:249)

Re: Nutch 2.0 error

Posted by Robert Irribarren <ro...@algorithms.io>.
I forgot to mention that my regex-urlfilter.txt looks like this

# skip file: ftp: and mailto: urls
-^(file|ftp|mailto):

# skip image and other suffixes we can't yet parse
# for a more extensive coverage use the urlfilter-suffix plugin
-\.(gif|GIF|jpg|JPG|png|PNG|ico|ICO|css|CSS|sit|SIT|eps|EPS|wmf|WMF|zip|ZIP|ppt|PPT|mpg|MPG|xls|XLS|gz|GZ|rpm|RPM|tgz|TGZ|mov|MOV|exe|EXE|jpeg|JPEG|bmp|BMP|js|JS)$

# skip URLs containing certain characters as probable queries, etc.
-[?*!@=]

# skip URLs with slash-delimited segment that repeats 3+ times, to break
loops
-.*(/[^/]+)/[^/]+\1/[^/]+\1/

# accept anything else
+.



On Mon, Aug 27, 2012 at 7:39 PM, Robert Irribarren <ro...@algorithms.io>wrote:

> I am running Nutch 2.0 in local mode with solr 4.0 beta
> I have this script here
>
> And I run this
> root@serverip:/usr/share/nutch/runtime/local# bin/nutch updatedb
> DbUpdaterJob: starting
> Exception in thread "main" java.lang.RuntimeException: job failed:
> name=update-table, jobid=job_local_0001
>         at
> org.apache.nutch.util.NutchJob.waitForCompletion(NutchJob.java:47)
>         at org.apache.nutch.crawl.DbUpdaterJob.run(DbUpdaterJob.java:96)
>         at
> org.apache.nutch.crawl.DbUpdaterJob.updateTable(DbUpdaterJob.java:103)
>         at org.apache.nutch.crawl.DbUpdaterJob.run(DbUpdaterJob.java:117)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>         at org.apache.nutch.crawl.DbUpdaterJob.main(DbUpdaterJob.java:121)
>
> I looked at the logs, but there wasn't much information here is the log
> GeneratorJob: Selecting best-scoring urls due for fetch.
> GeneratorJob: starting
> GeneratorJob: filtering: true
> GeneratorJob: topN: 18000
> GeneratorJob: done
> GeneratorJob: generated batch id: 1346109594-506815820
> FetcherJob: starting
> FetcherJob: batchId: 1346109594-506815820
> FetcherJob : timelimit set for : -1
> FetcherJob: threads: 10
> FetcherJob: parsing: false
> FetcherJob: resuming: false
> Using queue mode : byHost
> Fetcher: threads: 10
> QueueFeeder finished: total 0 records. Hit by time limit :0
> -finishing thread FetcherThread0, activeThreads=0
> Fetcher: throughput threshold: -1
> Fetcher: throughput threshold sequence: 5
> -finishing thread FetcherThread2, activeThreads=7
> -finishing thread FetcherThread3, activeThreads=6
> -finishing thread FetcherThread4, activeThreads=5
> -finishing thread FetcherThread5, activeThreads=4
> -finishing thread FetcherThread6, activeThreads=3
> -finishing thread FetcherThread7, activeThreads=2
> -finishing thread FetcherThread1, activeThreads=1
> -finishing thread FetcherThread8, activeThreads=0
> -finishing thread FetcherThread9, activeThreads=0
> 0/0 spinwaiting/active, 0 pages, 0 errors, 0.0 0.0 pages/s, 0 0 kb/s, 0
> URLs in 0 queues
> -activeThreads=0
> FetcherJob: done
> ParserJob: starting
> ParserJob: resuming:    false
> ParserJob: forced reparse:      false
> ParserJob: batchId:     1346109594-506815820
> Skipping http://www.questacon.edu.au/; different batch id
> Skipping http://www.cbc.ca/; different batch id
> Skipping http://www.ecokids.ca/; different batch id
> Skipping http://www.texted.ca/; different batch id
> Skipping http://www.texted.ca/app/en/; different batch id
> Skipping http://www.911forkids.com/; different batch id
> Skipping http://www.abcmouse.com/; different batch id
> Skipping http://get.adobe.com/flashplayer; different batch id
> Skipping http://get.adobe.com/flashplayer/; different batch id
> Skipping http://get.adobe.com/flashplayer/otherversions/; different batch
> id
> Skipping http://www.adobe.com/go/getflashplayer; different batch id
> Skipping http://www.afrigadget.com/; different batch id
> Skipping http://www.anamalz.com/; different batch id
> Skipping http://www.angelinaballerina.com/; different batch id
> Skipping http://www.angelinaballerina.com/usa/index.html; different batch
> id
> Skipping http://www.animaljam.com/; different batch id
> Skipping http://kids.aol.com/; different batch id
> Skipping http://www.aquariumofthebay.com/; different batch id
> Skipping http://www.bbc.com/news/; different batch id
> Skipping http://www.bbc.com/sport/; different batch id
> Skipping http://www.bbc.com/travel; different batch id
> Skipping http://www.bbc.com/travel/; different batch id
> Skipping http://www.bbcamerica.com/; different batch id
> Skipping
> http://www.bbcamericashop.com/dvd/life-discovery-channel-version-15686.html;
> different batch id
> Skipping http://bbcearth.com/; different batch id
> Skipping http://bbcearth.com/meet-your-planet; different batch id
> Skipping http://bbcearth.com/people; different batch id
> Skipping http://bbcearth.com/people/alastair-fothergill; different batch
> id
> Skipping http://www.bbc.co.uk/news/world_radio_and_tv/; different batch id
> Skipping http://www.bbc.co.uk/sport/0/; different batch id
> Skipping http://www.themouseclub.co.uk/; different batch id
> ParserJob: success
> DbUpdaterJob: starting
> SolrIndexerJob: starting
> SolrIndexerJob: done.
>
>
> The script I am running is
>
> #!/bin/bash
>
> # Nutch crawl
>
> export NUTCH_HOME=/usr/share/nutch/runtime/local
>
> # depth in the web exploration
> n=5
> # number of selected urls for fetching
> maxUrls=18000
> # solr server
> solrUrl=http://localhost:8983/solr/sites
>
>
> for (( i = 1 ; i <= $n ; i++ ))
> do
>
> log=$NUTCH_HOME/logs/log
>
> # Generate
> $NUTCH_HOME/bin/nutch generate -topN $maxUrls > $log
>
> batchId=`sed -n 's|.*batch id: \(.*\)|\1|p' < $log`
>
> # rename log file by appending the batch id
> log2=$log$batchId
> mv $log $log2
> log=$log2
>
> echo "Starting cycle $i of $n  Log file : $log2"
> # Fetch
> $NUTCH_HOME/bin/nutch fetch $batchId >> $log
>
> # Parse
> $NUTCH_HOME/bin/nutch parse $batchId >> $log
>
> # Update
> $NUTCH_HOME/bin/nutch updatedb >> $log
>
> # Index
> $NUTCH_HOME/bin/nutch solrindex $solrUrl $batchId >> $log
>
> done
>
> echo "starting finish crawl";
> bin/nutch parse -force -all
> bin/nutch updatedb
> bin/nutch solrindex http://127.0.0.1:8983/solr/sites -reindex
> echo "done"
>
>
>
> ---------------------------------
>
> It seems that when I change the
> # depth in the web exploration
> n=5
> # number of selected urls for fetching
> maxUrls=18000
>
> MaxURLS to anything higher, I don't get any newer results
> and The n=5 doesn't affect it as well, because each iteration it errors on
> the updatedb each time
> it outputs the same 3 lines
> adding 250 documents
> adding 250 documents
> adding 3 documents
>
> And my seeds have total up to 300 and I am running this on 618 MB of ram
> on the amazon ec2 free servers.
> Please Help!
>
>
>
> On Mon, Aug 27, 2012 at 5:09 AM, Lewis John Mcgibbney <
> lewis.mcgibbney@gmail.com> wrote:
>
>> Hi Robert,
>>
>> Please describe your problem and we will be more than happy to give
>> you a hand. The Nutch community is pretty active and in a very healthy
>> state, if people do not get back to your messages immediately then
>> don't be dissapointed, its because people  have lives outside of the
>> ASF and Nutch ;0)
>>
>> What version are you using, what version of Solr as well?
>> Do Solrj libraries match? All the usual stuff... lets try and debug
>> and get to the bottom of your error.
>>
>> Thanks
>>
>> Lewis
>>
>> On Sun, Aug 26, 2012 at 9:41 PM, Robert Irribarren <ro...@algorithms.io>
>> wrote:
>> > Thank you I just sent a few good ones, and I was fed up with no replys
>> so I
>> > just sent an error log with no description to see if people actually
>> cared.
>> > Thanks lewis for your response even if it shows no interest in the error
>> > itself but rather a course of action that I can follow to fit into the
>> > mailing list better. I thank you.
>> >
>> > On Sun, Aug 26, 2012 at 3:39 AM, Lewis John Mcgibbney <
>> > lewis.mcgibbney@gmail.com> wrote:
>> >
>> >> Hi Robert,,
>> >>
>> >>
>> >> On Sun, Aug 26, 2012 at 5:25 AM, Robert Irribarren <
>> robert@algorithms.io>
>> >> wrote:
>> >> > org.apache.solr.common.SolrException: Server Error
>> >> >
>> >> > Server Error
>> >> ...
>> >>
>> >> Please read this [0] before posting to the list. It saves both you and
>> >> us loads of time and also means there is less unnecessary noise of the
>> >> list.
>> >>
>> >> Thank you
>> >>
>> >> Lewis
>> >>
>> >> [0]
>> >>
>> http://wiki.apache.org/nutch/Becoming_A_Nutch_Developer#Step_One:_Using_the_Mailing_Lists
>> >>
>>
>>
>>
>> --
>> Lewis
>>
>
>

Re: Nutch 2.0 error

Posted by Robert Irribarren <ro...@algorithms.io>.
I am running Nutch 2.0 in local mode with solr 4.0 beta
I have this script here

And I run this
root@serverip:/usr/share/nutch/runtime/local# bin/nutch updatedb
DbUpdaterJob: starting
Exception in thread "main" java.lang.RuntimeException: job failed:
name=update-table, jobid=job_local_0001
        at
org.apache.nutch.util.NutchJob.waitForCompletion(NutchJob.java:47)
        at org.apache.nutch.crawl.DbUpdaterJob.run(DbUpdaterJob.java:96)
        at
org.apache.nutch.crawl.DbUpdaterJob.updateTable(DbUpdaterJob.java:103)
        at org.apache.nutch.crawl.DbUpdaterJob.run(DbUpdaterJob.java:117)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
        at org.apache.nutch.crawl.DbUpdaterJob.main(DbUpdaterJob.java:121)

I looked at the logs, but there wasn't much information here is the log
GeneratorJob: Selecting best-scoring urls due for fetch.
GeneratorJob: starting
GeneratorJob: filtering: true
GeneratorJob: topN: 18000
GeneratorJob: done
GeneratorJob: generated batch id: 1346109594-506815820
FetcherJob: starting
FetcherJob: batchId: 1346109594-506815820
FetcherJob : timelimit set for : -1
FetcherJob: threads: 10
FetcherJob: parsing: false
FetcherJob: resuming: false
Using queue mode : byHost
Fetcher: threads: 10
QueueFeeder finished: total 0 records. Hit by time limit :0
-finishing thread FetcherThread0, activeThreads=0
Fetcher: throughput threshold: -1
Fetcher: throughput threshold sequence: 5
-finishing thread FetcherThread2, activeThreads=7
-finishing thread FetcherThread3, activeThreads=6
-finishing thread FetcherThread4, activeThreads=5
-finishing thread FetcherThread5, activeThreads=4
-finishing thread FetcherThread6, activeThreads=3
-finishing thread FetcherThread7, activeThreads=2
-finishing thread FetcherThread1, activeThreads=1
-finishing thread FetcherThread8, activeThreads=0
-finishing thread FetcherThread9, activeThreads=0
0/0 spinwaiting/active, 0 pages, 0 errors, 0.0 0.0 pages/s, 0 0 kb/s, 0
URLs in 0 queues
-activeThreads=0
FetcherJob: done
ParserJob: starting
ParserJob: resuming:    false
ParserJob: forced reparse:      false
ParserJob: batchId:     1346109594-506815820
Skipping http://www.questacon.edu.au/; different batch id
Skipping http://www.cbc.ca/; different batch id
Skipping http://www.ecokids.ca/; different batch id
Skipping http://www.texted.ca/; different batch id
Skipping http://www.texted.ca/app/en/; different batch id
Skipping http://www.911forkids.com/; different batch id
Skipping http://www.abcmouse.com/; different batch id
Skipping http://get.adobe.com/flashplayer; different batch id
Skipping http://get.adobe.com/flashplayer/; different batch id
Skipping http://get.adobe.com/flashplayer/otherversions/; different batch id
Skipping http://www.adobe.com/go/getflashplayer; different batch id
Skipping http://www.afrigadget.com/; different batch id
Skipping http://www.anamalz.com/; different batch id
Skipping http://www.angelinaballerina.com/; different batch id
Skipping http://www.angelinaballerina.com/usa/index.html; different batch id
Skipping http://www.animaljam.com/; different batch id
Skipping http://kids.aol.com/; different batch id
Skipping http://www.aquariumofthebay.com/; different batch id
Skipping http://www.bbc.com/news/; different batch id
Skipping http://www.bbc.com/sport/; different batch id
Skipping http://www.bbc.com/travel; different batch id
Skipping http://www.bbc.com/travel/; different batch id
Skipping http://www.bbcamerica.com/; different batch id
Skipping
http://www.bbcamericashop.com/dvd/life-discovery-channel-version-15686.html;
different batch id
Skipping http://bbcearth.com/; different batch id
Skipping http://bbcearth.com/meet-your-planet; different batch id
Skipping http://bbcearth.com/people; different batch id
Skipping http://bbcearth.com/people/alastair-fothergill; different batch id
Skipping http://www.bbc.co.uk/news/world_radio_and_tv/; different batch id
Skipping http://www.bbc.co.uk/sport/0/; different batch id
Skipping http://www.themouseclub.co.uk/; different batch id
ParserJob: success
DbUpdaterJob: starting
SolrIndexerJob: starting
SolrIndexerJob: done.


The script I am running is

#!/bin/bash

# Nutch crawl

export NUTCH_HOME=/usr/share/nutch/runtime/local

# depth in the web exploration
n=5
# number of selected urls for fetching
maxUrls=18000
# solr server
solrUrl=http://localhost:8983/solr/sites


for (( i = 1 ; i <= $n ; i++ ))
do

log=$NUTCH_HOME/logs/log

# Generate
$NUTCH_HOME/bin/nutch generate -topN $maxUrls > $log

batchId=`sed -n 's|.*batch id: \(.*\)|\1|p' < $log`

# rename log file by appending the batch id
log2=$log$batchId
mv $log $log2
log=$log2

echo "Starting cycle $i of $n  Log file : $log2"
# Fetch
$NUTCH_HOME/bin/nutch fetch $batchId >> $log

# Parse
$NUTCH_HOME/bin/nutch parse $batchId >> $log

# Update
$NUTCH_HOME/bin/nutch updatedb >> $log

# Index
$NUTCH_HOME/bin/nutch solrindex $solrUrl $batchId >> $log

done

echo "starting finish crawl";
bin/nutch parse -force -all
bin/nutch updatedb
bin/nutch solrindex http://127.0.0.1:8983/solr/sites -reindex
echo "done"



---------------------------------

It seems that when I change the
# depth in the web exploration
n=5
# number of selected urls for fetching
maxUrls=18000

MaxURLS to anything higher, I don't get any newer results
and The n=5 doesn't affect it as well, because each iteration it errors on
the updatedb each time
it outputs the same 3 lines
adding 250 documents
adding 250 documents
adding 3 documents

And my seeds have total up to 300 and I am running this on 618 MB of ram on
the amazon ec2 free servers.
Please Help!


On Mon, Aug 27, 2012 at 5:09 AM, Lewis John Mcgibbney <
lewis.mcgibbney@gmail.com> wrote:

> Hi Robert,
>
> Please describe your problem and we will be more than happy to give
> you a hand. The Nutch community is pretty active and in a very healthy
> state, if people do not get back to your messages immediately then
> don't be dissapointed, its because people  have lives outside of the
> ASF and Nutch ;0)
>
> What version are you using, what version of Solr as well?
> Do Solrj libraries match? All the usual stuff... lets try and debug
> and get to the bottom of your error.
>
> Thanks
>
> Lewis
>
> On Sun, Aug 26, 2012 at 9:41 PM, Robert Irribarren <ro...@algorithms.io>
> wrote:
> > Thank you I just sent a few good ones, and I was fed up with no replys
> so I
> > just sent an error log with no description to see if people actually
> cared.
> > Thanks lewis for your response even if it shows no interest in the error
> > itself but rather a course of action that I can follow to fit into the
> > mailing list better. I thank you.
> >
> > On Sun, Aug 26, 2012 at 3:39 AM, Lewis John Mcgibbney <
> > lewis.mcgibbney@gmail.com> wrote:
> >
> >> Hi Robert,,
> >>
> >>
> >> On Sun, Aug 26, 2012 at 5:25 AM, Robert Irribarren <
> robert@algorithms.io>
> >> wrote:
> >> > org.apache.solr.common.SolrException: Server Error
> >> >
> >> > Server Error
> >> ...
> >>
> >> Please read this [0] before posting to the list. It saves both you and
> >> us loads of time and also means there is less unnecessary noise of the
> >> list.
> >>
> >> Thank you
> >>
> >> Lewis
> >>
> >> [0]
> >>
> http://wiki.apache.org/nutch/Becoming_A_Nutch_Developer#Step_One:_Using_the_Mailing_Lists
> >>
>
>
>
> --
> Lewis
>

Re: Nutch 2.0 error

Posted by Lewis John Mcgibbney <le...@gmail.com>.
Hi Robert,

Please describe your problem and we will be more than happy to give
you a hand. The Nutch community is pretty active and in a very healthy
state, if people do not get back to your messages immediately then
don't be dissapointed, its because people  have lives outside of the
ASF and Nutch ;0)

What version are you using, what version of Solr as well?
Do Solrj libraries match? All the usual stuff... lets try and debug
and get to the bottom of your error.

Thanks

Lewis

On Sun, Aug 26, 2012 at 9:41 PM, Robert Irribarren <ro...@algorithms.io> wrote:
> Thank you I just sent a few good ones, and I was fed up with no replys so I
> just sent an error log with no description to see if people actually cared.
> Thanks lewis for your response even if it shows no interest in the error
> itself but rather a course of action that I can follow to fit into the
> mailing list better. I thank you.
>
> On Sun, Aug 26, 2012 at 3:39 AM, Lewis John Mcgibbney <
> lewis.mcgibbney@gmail.com> wrote:
>
>> Hi Robert,,
>>
>>
>> On Sun, Aug 26, 2012 at 5:25 AM, Robert Irribarren <ro...@algorithms.io>
>> wrote:
>> > org.apache.solr.common.SolrException: Server Error
>> >
>> > Server Error
>> ...
>>
>> Please read this [0] before posting to the list. It saves both you and
>> us loads of time and also means there is less unnecessary noise of the
>> list.
>>
>> Thank you
>>
>> Lewis
>>
>> [0]
>> http://wiki.apache.org/nutch/Becoming_A_Nutch_Developer#Step_One:_Using_the_Mailing_Lists
>>



-- 
Lewis

Re: Nutch 2.0 error

Posted by Robert Irribarren <ro...@algorithms.io>.
Thank you I just sent a few good ones, and I was fed up with no replys so I
just sent an error log with no description to see if people actually cared.
Thanks lewis for your response even if it shows no interest in the error
itself but rather a course of action that I can follow to fit into the
mailing list better. I thank you.

On Sun, Aug 26, 2012 at 3:39 AM, Lewis John Mcgibbney <
lewis.mcgibbney@gmail.com> wrote:

> Hi Robert,,
>
>
> On Sun, Aug 26, 2012 at 5:25 AM, Robert Irribarren <ro...@algorithms.io>
> wrote:
> > org.apache.solr.common.SolrException: Server Error
> >
> > Server Error
> ...
>
> Please read this [0] before posting to the list. It saves both you and
> us loads of time and also means there is less unnecessary noise of the
> list.
>
> Thank you
>
> Lewis
>
> [0]
> http://wiki.apache.org/nutch/Becoming_A_Nutch_Developer#Step_One:_Using_the_Mailing_Lists
>

Re: Nutch 2.0 error

Posted by Lewis John Mcgibbney <le...@gmail.com>.
Hi Robert,,


On Sun, Aug 26, 2012 at 5:25 AM, Robert Irribarren <ro...@algorithms.io> wrote:
> org.apache.solr.common.SolrException: Server Error
>
> Server Error
...

Please read this [0] before posting to the list. It saves both you and
us loads of time and also means there is less unnecessary noise of the
list.

Thank you

Lewis

[0] http://wiki.apache.org/nutch/Becoming_A_Nutch_Developer#Step_One:_Using_the_Mailing_Lists