You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by "Kevin Risden (JIRA)" <ji...@apache.org> on 2016/10/03 20:07:23 UTC

[jira] [Commented] (SOLR-9591) Shards and replicas go down when indexing large number of files

    [ https://issues.apache.org/jira/browse/SOLR-9591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15543274#comment-15543274 ] 

Kevin Risden commented on SOLR-9591:
------------------------------------

Can you provide the full stack trace? Redact anything sensitive. If the whole log isn't too big that could be helpful too.

Few other questions:
* Are these plain text files or something else?
* Does it happen with any set of text files you tried?
* Are these text files publically available?
* Since you trying single node with embedded, does this same error occur with the latest Solr release?

> Shards and replicas go down when indexing large number of files
> ---------------------------------------------------------------
>
>                 Key: SOLR-9591
>                 URL: https://issues.apache.org/jira/browse/SOLR-9591
>             Project: Solr
>          Issue Type: Bug
>      Security Level: Public(Default Security Level. Issues are Public) 
>          Components: SolrCloud
>    Affects Versions: 5.5.2
>            Reporter: Khalid Alharbi
>
> Solr shards and replicas go down when indexing a large number of text files using the default [extracting request handler|https://cwiki.apache.org/confluence/x/c4DxAQ].
> {code}
> curl 'http://localhost:8983/solr/myCollection/update/extract?literal.id=someId' -F "myfile=/data/file1.txt"
> {code}
> and committing after indexing 5,000 files using:
> {code}
> curl 'http://localhost:8983/solr/myCollection/update?commit=true&wt=json'
> {code}
> This was on Solr (SolrCloud) version 5.5.2 with an external zookeeper cluster 
> of five nodes. I also tried this on a single node SolrCloud with the embedded ZooKeeper but the collection went down as well. In both cases the error message is always "ERROR null DistributedUpdateProcessor ClusterState says we are the leader,​ but locally we don't think so"
> I managed to come up with a work around that helped me index over 400K files without getting replicas down with that error message. The work around is to index 5K files, restart Solr, wait for shards and replicas to get active, then index the next 5K files, and repeat the previous steps.
> If this is not enough to investigate this issue, I will be happy to provide more details regarding this issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@lucene.apache.org
For additional commands, e-mail: dev-help@lucene.apache.org