You are viewing a plain text version of this content. The canonical link for it is here.
Posted to solr-user@lucene.apache.org by "atuldj.jadhav" <at...@gmail.com> on 2013/09/09 00:01:43 UTC

SOLR index Recovery & availability

Hi Team,Need your suggestions/views on the approach I have in place for SOLR
availability and recovery.
I am running *SOLR 3.5* and have around *30k* document's indexed in my SOLR
core. I have configured SOLR to hold *5k * documents in each segment at a
time.I periodically commit & optimize my SOLR index. 

I have delta indexing in place to index new documents in SOLR, /very rarely
/ I face index corruption issue, to fix this issue I have *checkindex -fix*
job in place as well.However sometime this job can delete the corrupt
segment! (meaning loss of 5K documents, till I full Re-index SOLR.)

_*I have few follow up questions on this case.*_
1. How can I avoid loss of 5K documents (checkindex -fix), shall I reduce
number of documents per segments count? is there an alternate solution?

2. If I start taking periodic backup (snapshot) of entire index, shall I
just replace my data/index folder from the backup folder in case corruption
is found? Is this a good implementation? 

3. Any other good solution, suggestion to have maximum index availability
all the time? 

Thanks in advance for giving your time. 

Atul 



--
View this message in context: http://lucene.472066.n3.nabble.com/SOLR-index-Recovery-availability-tp4088782.html
Sent from the Solr - User mailing list archive at Nabble.com.

Re: SOLR index Recovery & availability

Posted by Walter Underwood <wu...@wunderwood.org>.
This sounds very complicated for only 30K documents. Put them all on one server, give it enough memory so that the index can all be in file buffers. If there is a disaster, reindex everything. That should only take a few minutes.

And don't optimize.

wunder

On Sep 8, 2013, at 3:01 PM, atuldj.jadhav wrote:

> Hi Team,Need your suggestions/views on the approach I have in place for SOLR
> availability and recovery.
> I am running *SOLR 3.5* and have around *30k* document's indexed in my SOLR
> core. I have configured SOLR to hold *5k * documents in each segment at a
> time.I periodically commit & optimize my SOLR index. 
> 
> I have delta indexing in place to index new documents in SOLR, /very rarely
> / I face index corruption issue, to fix this issue I have *checkindex -fix*
> job in place as well.However sometime this job can delete the corrupt
> segment! (meaning loss of 5K documents, till I full Re-index SOLR.)
> 
> _*I have few follow up questions on this case.*_
> 1. How can I avoid loss of 5K documents (checkindex -fix), shall I reduce
> number of documents per segments count? is there an alternate solution?
> 
> 2. If I start taking periodic backup (snapshot) of entire index, shall I
> just replace my data/index folder from the backup folder in case corruption
> is found? Is this a good implementation? 
> 
> 3. Any other good solution, suggestion to have maximum index availability
> all the time? 
> 
> Thanks in advance for giving your time. 
> 
> Atul 
> 
> 
> 
> --
> View this message in context: http://lucene.472066.n3.nabble.com/SOLR-index-Recovery-availability-tp4088782.html
> Sent from the Solr - User mailing list archive at Nabble.com.

--
Walter Underwood
wunder@wunderwood.org