You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by "Fredrik Rodland (JIRA)" <ji...@apache.org> on 2014/11/06 10:01:34 UTC

[jira] [Comment Edited] (SOLR-4587) Implement Saved Searches a la ElasticSearch Percolator

    [ https://issues.apache.org/jira/browse/SOLR-4587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14200006#comment-14200006 ] 

Fredrik Rodland edited comment on SOLR-4587 at 11/6/14 9:01 AM:
----------------------------------------------------------------

Sound good!

Having implemented a pretty large system for matching documents against queries (using elasticsearch to index the queries) we discovered very early that filtering the queries was an important requirement to get things running with acceptable performance. 

So I would add to your list of acceptance criteria that the request must support *fq* and that this is performed prior to the looping.  This would enable us to get a smaller list of queries to loop and thus reducing the time to complete the request.  For this to work queries also need to support filter-fields - i.e. regular solr fields in addition to the fq, q, defType, etc mentioned above.

For the record our system has ≈1mill queries, and we're matching ≈10 doc/s.  I believe that much of the job in luwak also comes from the realization that the number of documents must be reduced prior to looping.  I'm sure [~romseygeek] can elaborate on this as well.


was (Author: fmr):
Sound good!

Having implemented a pretty large system for matching documents against queries (using elasticsearch to index the queries) we discovered very early that filtering the queries was an important requirement to get things running with acceptable performance. 

So I would add to your list of acceptance criteria that the request must support *fq* and that this is performed prior to the looping.  This would enable us to get a smaller list of queries to loop and thus reducing the time to complete the request.  For this to work queries also need to support filter-fields - i.e. regular solr fields in addition to the fq, q, defType, etc mentioned above.

For the record our system has ≈1mill queries, and we're matching ≈10 doc/s.  I believe that much of the job in luwak also comes from the realization that the number of filters must be reduced prior to looping.  I'm sure [~romseygeek] can elaborate on this as well.

> Implement Saved Searches a la ElasticSearch Percolator
> ------------------------------------------------------
>
>                 Key: SOLR-4587
>                 URL: https://issues.apache.org/jira/browse/SOLR-4587
>             Project: Solr
>          Issue Type: New Feature
>          Components: SearchComponents - other, SolrCloud
>            Reporter: Otis Gospodnetic
>             Fix For: Trunk
>
>
> Use Lucene MemoryIndex for this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@lucene.apache.org
For additional commands, e-mail: dev-help@lucene.apache.org