You are viewing a plain text version of this content. The canonical link for it is here.
Posted to solr-user@lucene.apache.org by Vadim Kisselmann <v....@googlemail.com> on 2011/08/04 14:30:11 UTC

Unbuffered entity enclosing request can not be repeated & Invalid chunk header

Hello folks,

i use solr 1.4.1 and every 2 to 6 hours i have indexing errors in my log
files.

on the client side:
2011-08-04 12:01:18,966 ERROR [Worker-242] IndexServiceImpl - Indexing
failed with SolrServerException.
Details: org.apache.commons.httpclient.ProtocolException: Unbuffered entity
enclosing request can not be repeated.:
Stacktrace: org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:469)
.
.
on the server side:
INFO: [] webapp=/solr path=/update params={wt=javabin&version=1} status=0
QTime=3
04.08.2011 12:01:18 org.apache.solr.update.processor.LogUpdateProcessor
finish
INFO: {} 0 0
04.08.2011 12:01:18 org.apache.solr.common.SolrException log
SCHWERWIEGEND: org.apache.solr.common.SolrException: java.io.IOException:
Invalid chunk header
.
.
.
i`m indexing ONE document per call, 15-20 documents per second, 24/7.
what may be the problem?

best regards
vadim

Re: Unbuffered entity enclosing request can not be repeated & Invalid chunk header

Posted by Markus Jelsma <ma...@openindex.io>.
Hi,

> Hi Markus,
> 
> thanks for your answer.
> I'm using Solr. 4.0 and jetty now and observe the behavior and my error
> logs next week.
> tomcat can be a reason, we will see, i'll report.
> 
> I'm indexing WITHOUT batches, one doc after another. But i would try out
> the batch indexing as well as
> retry indexing faulty docs.
> if you indexing one batch, and one doc in batch is corrupt, what happens
> with another 249docs(total 250/batch)? Are they indexed and
> updated when you retry to indexing the batch, or fails the complete batch?

The entire batch should fail but i cannot confirm. Usually all fail if there 
is an error somewhere such as an XML error.

> 
> Regards
> Vadim
> 
> 
> 
> 
> 2011/8/11 Markus Jelsma <ma...@openindex.io>
> 
> > Hi,
> > 
> > We  see these errors too once on a while but there is real answer on the
> > mailing list here except one user suspecting Tomcat is responsible
> > (connection
> > time outs).
> > 
> > Another user proposed to limit the number of documents per batch but
> > that, of
> > course, increases the number of connections made. We do only 250
> > docs/batch to
> > limit RAM usage on the client and start to see these errors very
> > occasionally.
> > There may be a coincidence.. or not.
> > 
> > Anyway, it's really hard to reproduce if not impossible. It happens when
> > connecting directly as well when connecting through a proxy.
> > 
> > What you can do is simply retry the batch and it usually works out fine.
> > At least you don't loose a batch in the process. We retry all failures
> > at least a
> > couple of times before giving up an indexing job.
> > 
> > Cheers,
> > 
> > > Hello folks,
> > > 
> > > i use solr 1.4.1 and every 2 to 6 hours i have indexing errors in my
> > > log files.
> > > 
> > > on the client side:
> > > 2011-08-04 12:01:18,966 ERROR [Worker-242] IndexServiceImpl - Indexing
> > > failed with SolrServerException.
> > > Details: org.apache.commons.httpclient.ProtocolException: Unbuffered
> > 
> > entity
> > 
> > > enclosing request can not be repeated.:
> > 
> > > Stacktrace:
> > org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHt
> > t
> > 
> > > pSolrServer.java:469) .
> > > .
> > > on the server side:
> > > INFO: [] webapp=/solr path=/update params={wt=javabin&version=1}
> > > status=0 QTime=3
> > > 04.08.2011 12:01:18 org.apache.solr.update.processor.LogUpdateProcessor
> > > finish
> > > INFO: {} 0 0
> > > 04.08.2011 12:01:18 org.apache.solr.common.SolrException log
> > > SCHWERWIEGEND: org.apache.solr.common.SolrException:
> > > java.io.IOException: Invalid chunk header
> > > .
> > > .
> > > .
> > > i`m indexing ONE document per call, 15-20 documents per second, 24/7.
> > > what may be the problem?
> > > 
> > > best regards
> > > vadim

Re: Unbuffered entity enclosing request can not be repeated & Invalid chunk header

Posted by Vadim Kisselmann <v....@googlemail.com>.
Hi Markus,

thanks for your answer.
I'm using Solr. 4.0 and jetty now and observe the behavior and my error logs
next week.
tomcat can be a reason, we will see, i'll report.

I'm indexing WITHOUT batches, one doc after another. But i would try out the
batch indexing as well as
retry indexing faulty docs.
if you indexing one batch, and one doc in batch is corrupt, what happens
with another 249docs(total 250/batch)? Are they indexed and
updated when you retry to indexing the batch, or fails the complete batch?

Regards
Vadim




2011/8/11 Markus Jelsma <ma...@openindex.io>

> Hi,
>
> We  see these errors too once on a while but there is real answer on the
> mailing list here except one user suspecting Tomcat is responsible
> (connection
> time outs).
>
> Another user proposed to limit the number of documents per batch but that,
> of
> course, increases the number of connections made. We do only 250 docs/batch
> to
> limit RAM usage on the client and start to see these errors very
> occasionally.
> There may be a coincidence.. or not.
>
> Anyway, it's really hard to reproduce if not impossible. It happens when
> connecting directly as well when connecting through a proxy.
>
> What you can do is simply retry the batch and it usually works out fine. At
> least you don't loose a batch in the process. We retry all failures at
> least a
> couple of times before giving up an indexing job.
>
> Cheers,
>
> > Hello folks,
> >
> > i use solr 1.4.1 and every 2 to 6 hours i have indexing errors in my log
> > files.
> >
> > on the client side:
> > 2011-08-04 12:01:18,966 ERROR [Worker-242] IndexServiceImpl - Indexing
> > failed with SolrServerException.
> > Details: org.apache.commons.httpclient.ProtocolException: Unbuffered
> entity
> > enclosing request can not be repeated.:
> > Stacktrace:
> >
> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHtt
> > pSolrServer.java:469) .
> > .
> > on the server side:
> > INFO: [] webapp=/solr path=/update params={wt=javabin&version=1} status=0
> > QTime=3
> > 04.08.2011 12:01:18 org.apache.solr.update.processor.LogUpdateProcessor
> > finish
> > INFO: {} 0 0
> > 04.08.2011 12:01:18 org.apache.solr.common.SolrException log
> > SCHWERWIEGEND: org.apache.solr.common.SolrException: java.io.IOException:
> > Invalid chunk header
> > .
> > .
> > .
> > i`m indexing ONE document per call, 15-20 documents per second, 24/7.
> > what may be the problem?
> >
> > best regards
> > vadim
>

Re: Unbuffered entity enclosing request can not be repeated & Invalid chunk header

Posted by Markus Jelsma <ma...@openindex.io>.
Hi,

We  see these errors too once on a while but there is real answer on the 
mailing list here except one user suspecting Tomcat is responsible (connection 
time outs).

Another user proposed to limit the number of documents per batch but that, of 
course, increases the number of connections made. We do only 250 docs/batch to 
limit RAM usage on the client and start to see these errors very occasionally. 
There may be a coincidence.. or not. 

Anyway, it's really hard to reproduce if not impossible. It happens when 
connecting directly as well when connecting through a proxy.

What you can do is simply retry the batch and it usually works out fine. At 
least you don't loose a batch in the process. We retry all failures at least a 
couple of times before giving up an indexing job.

Cheers,

> Hello folks,
> 
> i use solr 1.4.1 and every 2 to 6 hours i have indexing errors in my log
> files.
> 
> on the client side:
> 2011-08-04 12:01:18,966 ERROR [Worker-242] IndexServiceImpl - Indexing
> failed with SolrServerException.
> Details: org.apache.commons.httpclient.ProtocolException: Unbuffered entity
> enclosing request can not be repeated.:
> Stacktrace:
> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHtt
> pSolrServer.java:469) .
> .
> on the server side:
> INFO: [] webapp=/solr path=/update params={wt=javabin&version=1} status=0
> QTime=3
> 04.08.2011 12:01:18 org.apache.solr.update.processor.LogUpdateProcessor
> finish
> INFO: {} 0 0
> 04.08.2011 12:01:18 org.apache.solr.common.SolrException log
> SCHWERWIEGEND: org.apache.solr.common.SolrException: java.io.IOException:
> Invalid chunk header
> .
> .
> .
> i`m indexing ONE document per call, 15-20 documents per second, 24/7.
> what may be the problem?
> 
> best regards
> vadim