You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@couchdb.apache.org by Ramkrishna Kulkarni <ra...@gmail.com> on 2011/07/22 13:39:13 UTC

CouchDB filters: {"error":"changes_reader_died"}

I have a DB with some 40K documents. I initiate around 40 replications at
the same time with a filter. The filter is a simple string comparison on one
of the fields. On an average, each replication copies 1K documents to the
target DBs. This process takes several minutes (30 minutes sometimes). I
have observed that around 90% of the time is spent in filtering the
documents at source DB (I'm guessing this because CPU is fully loaded at
source for most of the time and once the copying starts, it finishes pretty
quickly). Situation is a little better if the number of simultaneously
replications is less.

This DB has 4 views (composite keys). Tried on 1.0.2 and 1.1.0. Server is a
2 core box running Ubuntu 10.10.
I've seen message like "changes_timeout" and {"error":"changes_reader_**
died"}

Please let me know if there are things to keep in mind while using filters.

Thanks,
Ram

Re: CouchDB filters: {"error":"changes_reader_died"}

Posted by Hans-Dieter Böhlau <bo...@googlemail.com>.
Hi, we run into the same problems: [error] [<0.10873.5>] changes loop
timeout, no data received

We want periodically to replicate a subset of documents into a new database.
The number of documents in the source-database grows over time. As we
noticed, the time until the changes request provides a response increases
too. In some cases it runs in into a timeout.

Please can anyone explain, why filtered replication seems to be so
expensive?!
Does the reponse time depend on the number of documents, the sequence
numbers. Or has the server load a significant impact?

Think, https://issues.apache.org/jira/browse/COUCHDB-1231 points on this
issue.

Best regards,
Hans


2011/7/22 Ramkrishna Kulkarni <ra...@gmail.com>

> I have a DB with some 40K documents. I initiate around 40 replications at
> the same time with a filter. The filter is a simple string comparison on
> one
> of the fields. On an average, each replication copies 1K documents to the
> target DBs. This process takes several minutes (30 minutes sometimes). I
> have observed that around 90% of the time is spent in filtering the
> documents at source DB (I'm guessing this because CPU is fully loaded at
> source for most of the time and once the copying starts, it finishes pretty
> quickly). Situation is a little better if the number of simultaneously
> replications is less.
>
> This DB has 4 views (composite keys). Tried on 1.0.2 and 1.1.0. Server is a
> 2 core box running Ubuntu 10.10.
> I've seen message like "changes_timeout" and {"error":"changes_reader_**
> died"}
>
> Please let me know if there are things to keep in mind while using filters.
>
> Thanks,
> Ram
>