You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by rm...@apache.org on 2011/10/01 05:05:07 UTC
svn commit: r1177888 [13/16] - in /lucene/dev/branches/lucene2621: ./
dev-tools/eclipse/ dev-tools/idea/lucene/contrib/ dev-tools/maven/ lucene/
lucene/contrib/ lucene/contrib/demo/src/java/org/apache/lucene/demo/
lucene/contrib/demo/src/java/org/apach...
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/core/SolrCore.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/core/SolrCore.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/core/SolrCore.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/core/SolrCore.java Sat Oct 1 03:04:53 2011
@@ -1193,78 +1193,122 @@ public final class SolrCore implements S
// warm the new searcher based on the current searcher.
// should this go before the other event handlers or after?
if (currSearcher != null) {
- future = searcherExecutor.submit(
- new Callable() {
- public Object call() throws Exception {
- try {
- newSearcher.warm(currSearcher);
- } catch (Throwable e) {
- SolrException.logOnce(log,null,e);
+ try {
+ future = searcherExecutor.submit(
+ new Callable() {
+ public Object call() throws Exception {
+ try {
+ newSearcher.warm(currSearcher);
+ } catch (Throwable e) {
+ SolrException.logOnce(log,null,e);
+ }
+ return null;
}
- return null;
}
- }
- );
+ );
+ } catch(Exception e) {
+ // if submit fails, newSearchHolder does not get decref'd
+ if (newSearchHolder != null) {
+ newSearchHolder.decref();
+ if (returnSearcher) {
+ newSearchHolder.decref();
+ }
+ }
+ throw e;
+ }
}
if (currSearcher==null && firstSearcherListeners.size() > 0) {
- future = searcherExecutor.submit(
- new Callable() {
- public Object call() throws Exception {
- try {
- for (SolrEventListener listener : firstSearcherListeners) {
- listener.newSearcher(newSearcher,null);
+ try {
+ future = searcherExecutor.submit(
+ new Callable() {
+ public Object call() throws Exception {
+ try {
+ for (SolrEventListener listener : firstSearcherListeners) {
+ listener.newSearcher(newSearcher,null);
+ }
+ } catch (Throwable e) {
+ SolrException.logOnce(log,null,e);
}
- } catch (Throwable e) {
- SolrException.logOnce(log,null,e);
+ return null;
}
- return null;
}
- }
- );
+ );
+ } catch(Exception e) {
+ // if submit fails, newSearchHolder does not get decref'd
+ if (newSearchHolder != null) {
+ newSearchHolder.decref();
+ if (returnSearcher) {
+ newSearchHolder.decref();
+ }
+ }
+ throw e;
+ }
}
if (currSearcher!=null && newSearcherListeners.size() > 0) {
- future = searcherExecutor.submit(
- new Callable() {
- public Object call() throws Exception {
- try {
- for (SolrEventListener listener : newSearcherListeners) {
- listener.newSearcher(newSearcher, currSearcher);
+ try {
+ future = searcherExecutor.submit(
+ new Callable() {
+ public Object call() throws Exception {
+ try {
+ for (SolrEventListener listener : newSearcherListeners) {
+ listener.newSearcher(newSearcher, currSearcher);
+ }
+ } catch (Throwable e) {
+ SolrException.logOnce(log,null,e);
}
- } catch (Throwable e) {
- SolrException.logOnce(log,null,e);
+ return null;
}
- return null;
}
- }
- );
+ );
+ } catch(Exception e) {
+ // if submit fails, newSearchHolder does not get decref'd
+ if (newSearchHolder != null) {
+ newSearchHolder.decref();
+ if (returnSearcher) {
+ newSearchHolder.decref();
+ }
+ }
+ throw e;
+ }
}
// WARNING: this code assumes a single threaded executor (that all tasks
// queued will finish first).
final RefCounted<SolrIndexSearcher> currSearcherHolderF = currSearcherHolder;
if (!alreadyRegistered) {
- future = searcherExecutor.submit(
- new Callable() {
- public Object call() throws Exception {
- try {
- // signal that we no longer need to decrement
- // the count *before* registering the searcher since
- // registerSearcher will decrement even if it errors.
- decrementOnDeckCount[0]=false;
- registerSearcher(newSearchHolder);
- } catch (Throwable e) {
- SolrException.logOnce(log,null,e);
- } finally {
- // we are all done with the old searcher we used
- // for warming...
- if (currSearcherHolderF!=null) currSearcherHolderF.decref();
+ try {
+ future = searcherExecutor.submit(
+ new Callable() {
+ public Object call() throws Exception {
+ try {
+ // signal that we no longer need to decrement
+ // the count *before* registering the searcher since
+ // registerSearcher will decrement even if it errors.
+ decrementOnDeckCount[0]=false;
+ registerSearcher(newSearchHolder);
+ } catch (Throwable e) {
+ SolrException.logOnce(log,null,e);
+ } finally {
+ // we are all done with the old searcher we used
+ // for warming...
+ if (currSearcherHolderF!=null) currSearcherHolderF.decref();
+ }
+ return null;
}
- return null;
}
- }
- );
+ );
+ } catch(Exception e) {
+ // if submit fails, newSearchHolder does not get decref'd
+ if (newSearchHolder != null) {
+ newSearchHolder.decref();
+ if (returnSearcher) {
+ newSearchHolder.decref();
+ }
+ }
+ throw e;
+ }
}
if (waitSearcher != null) {
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/AnalysisRequestHandlerBase.java Sat Oct 1 03:04:53 2011
@@ -88,7 +88,7 @@ public abstract class AnalysisRequestHan
TokenStream tokenStream = null;
try {
- tokenStream = analyzer.reusableTokenStream(context.getFieldName(), new StringReader(value));
+ tokenStream = analyzer.tokenStream(context.getFieldName(), new StringReader(value));
} catch (IOException e) {
throw new SolrException(SolrException.ErrorCode.BAD_REQUEST, e);
}
@@ -113,7 +113,7 @@ public abstract class AnalysisRequestHan
}
}
- TokenStream tokenStream = tfac.create(tokenizerChain.charStream(new StringReader(value)));
+ TokenStream tokenStream = tfac.create(tokenizerChain.initReader(new StringReader(value)));
List<AttributeSource> tokens = analyzeTokenStream(tokenStream);
namedList.add(tokenStream.getClass().getName(), convertTokensToNamedLists(tokens, context));
@@ -140,20 +140,25 @@ public abstract class AnalysisRequestHan
* @param analyzer The analyzer to use.
*/
protected Set<BytesRef> getQueryTokenSet(String query, Analyzer analyzer) {
- final Set<BytesRef> tokens = new HashSet<BytesRef>();
- final TokenStream tokenStream = analyzer.tokenStream("", new StringReader(query));
- final TermToBytesRefAttribute bytesAtt = tokenStream.getAttribute(TermToBytesRefAttribute.class);
- final BytesRef bytes = bytesAtt.getBytesRef();
try {
+ final Set<BytesRef> tokens = new HashSet<BytesRef>();
+ final TokenStream tokenStream = analyzer.tokenStream("", new StringReader(query));
+ final TermToBytesRefAttribute bytesAtt = tokenStream.getAttribute(TermToBytesRefAttribute.class);
+ final BytesRef bytes = bytesAtt.getBytesRef();
+
tokenStream.reset();
+
while (tokenStream.incrementToken()) {
bytesAtt.fillBytesRef();
tokens.add(new BytesRef(bytes));
}
+
+ tokenStream.end();
+ tokenStream.close();
+ return tokens;
} catch (IOException ioe) {
throw new RuntimeException("Error occured while iterating over tokenstream", ioe);
}
- return tokens;
}
/**
@@ -197,7 +202,7 @@ public abstract class AnalysisRequestHan
/**
* Converts the list of Tokens to a list of NamedLists representing the tokens.
*
- * @param tokens Tokens to convert
+ * @param tokenList Tokens to convert
* @param context The analysis context
*
* @return List of NamedLists containing the relevant information taken from the tokens
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java Sat Oct 1 03:04:53 2011
@@ -183,8 +183,9 @@ public class ReplicationHandler extends
rsp.add("message","No slave configured");
}
} else if (command.equalsIgnoreCase(CMD_ABORT_FETCH)) {
- if (snapPuller != null){
- snapPuller.abortPull();
+ SnapPuller temp = tempSnapPuller;
+ if (temp != null){
+ temp.abortPull();
rsp.add(STATUS, OK_STATUS);
} else {
rsp.add(STATUS,ERR_STATUS);
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java Sat Oct 1 03:04:53 2011
@@ -309,13 +309,15 @@ public class QueryElevationComponent ext
return query;
}
StringBuilder norm = new StringBuilder();
- TokenStream tokens = analyzer.reusableTokenStream( "", new StringReader( query ) );
+ TokenStream tokens = analyzer.tokenStream("", new StringReader(query));
tokens.reset();
CharTermAttribute termAtt = tokens.addAttribute(CharTermAttribute.class);
while( tokens.incrementToken() ) {
norm.append( termAtt.buffer(), 0, termAtt.length() );
}
+ tokens.end();
+ tokens.close();
return norm.toString();
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/SpellCheckComponent.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/SpellCheckComponent.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/SpellCheckComponent.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/handler/component/SpellCheckComponent.java Sat Oct 1 03:04:53 2011
@@ -474,7 +474,7 @@ public class SpellCheckComponent extends
private Collection<Token> getTokens(String q, Analyzer analyzer) throws IOException {
Collection<Token> result = new ArrayList<Token>();
assert analyzer != null;
- TokenStream ts = analyzer.reusableTokenStream("", new StringReader(q));
+ TokenStream ts = analyzer.tokenStream("", new StringReader(q));
ts.reset();
// TODO: support custom attributes
CharTermAttribute termAtt = ts.addAttribute(CharTermAttribute.class);
@@ -494,6 +494,8 @@ public class SpellCheckComponent extends
token.setPositionIncrement(posIncAtt.getPositionIncrement());
result.add(token);
}
+ ts.end();
+ ts.close();
return result;
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/highlight/DefaultSolrHighlighter.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/highlight/DefaultSolrHighlighter.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/highlight/DefaultSolrHighlighter.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/highlight/DefaultSolrHighlighter.java Sat Oct 1 03:04:53 2011
@@ -599,7 +599,7 @@ public class DefaultSolrHighlighter exte
private TokenStream createAnalyzerTStream(IndexSchema schema, String fieldName, String docText) throws IOException {
TokenStream tstream;
- TokenStream ts = schema.getAnalyzer().reusableTokenStream(fieldName, new StringReader(docText));
+ TokenStream ts = schema.getAnalyzer().tokenStream(fieldName, new StringReader(docText));
ts.reset();
tstream = new TokenOrderingFilter(ts, 10);
return tstream;
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/BoolField.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/BoolField.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/BoolField.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/BoolField.java Sat Oct 1 03:04:53 2011
@@ -30,7 +30,6 @@ import org.apache.lucene.util.CharsRef;
import org.apache.lucene.util.mutable.MutableValue;
import org.apache.lucene.util.mutable.MutableValueBool;
import org.apache.solr.search.QParser;
-import org.apache.solr.search.function.*;
import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.Tokenizer;
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
@@ -70,7 +69,7 @@ public class BoolField extends FieldType
protected final static Analyzer boolAnalyzer = new SolrAnalyzer() {
@Override
- public TokenStreamInfo getStream(String fieldName, Reader reader) {
+ public TokenStreamComponents createComponents(String fieldName, Reader reader) {
Tokenizer tokenizer = new Tokenizer(reader) {
final CharTermAttribute termAtt = addAttribute(CharTermAttribute.class);
boolean done = false;
@@ -95,7 +94,7 @@ public class BoolField extends FieldType
}
};
- return new TokenStreamInfo(tokenizer, tokenizer);
+ return new TokenStreamComponents(tokenizer);
}
};
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/CollationField.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/CollationField.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/CollationField.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/CollationField.java Sat Oct 1 03:04:53 2011
@@ -214,10 +214,10 @@ public class CollationField extends Fiel
TokenStream source;
try {
- source = analyzer.reusableTokenStream(field, new StringReader(part));
+ source = analyzer.tokenStream(field, new StringReader(part));
source.reset();
} catch (IOException e) {
- source = analyzer.tokenStream(field, new StringReader(part));
+ throw new RuntimeException("Unable to initialize TokenStream to analyze range part: " + part, e);
}
TermToBytesRefAttribute termAtt = source.getAttribute(TermToBytesRefAttribute.class);
@@ -234,8 +234,11 @@ public class CollationField extends Fiel
}
try {
+ source.end();
source.close();
- } catch (IOException ignored) {}
+ } catch (IOException e) {
+ throw new RuntimeException("Unable to end & close TokenStream after analyzing range part: " + part, e);
+ }
return new BytesRef(bytes);
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/FieldType.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/FieldType.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/FieldType.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/FieldType.java Sat Oct 1 03:04:53 2011
@@ -273,7 +273,7 @@ public abstract class FieldType extends
* @return the {@link org.apache.lucene.index.IndexableField}.
*/
protected IndexableField createField(String name, String val, org.apache.lucene.document.FieldType type, float boost){
- Field f = new Field(name, type, val);
+ Field f = new Field(name, val, type);
f.setBoost(boost);
return f;
}
@@ -389,7 +389,7 @@ public abstract class FieldType extends
}
@Override
- public TokenStreamInfo getStream(String fieldName, Reader reader) {
+ public TokenStreamComponents createComponents(String fieldName, Reader reader) {
Tokenizer ts = new Tokenizer(reader) {
final char[] cbuf = new char[maxChars];
final CharTermAttribute termAtt = addAttribute(CharTermAttribute.class);
@@ -406,7 +406,7 @@ public abstract class FieldType extends
}
};
- return new TokenStreamInfo(ts, ts);
+ return new TokenStreamComponents(ts);
}
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/IndexSchema.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/IndexSchema.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/IndexSchema.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/IndexSchema.java Sat Oct 1 03:04:53 2011
@@ -18,7 +18,7 @@
package org.apache.solr.schema;
import org.apache.lucene.analysis.Analyzer;
-import org.apache.lucene.analysis.TokenStream;
+import org.apache.lucene.analysis.AnalyzerWrapper;
import org.apache.lucene.index.IndexableField;
import org.apache.lucene.search.similarities.DefaultSimilarity;
import org.apache.lucene.search.similarities.Similarity;
@@ -41,8 +41,6 @@ import org.xml.sax.InputSource;
import javax.xml.xpath.XPath;
import javax.xml.xpath.XPathConstants;
import javax.xml.xpath.XPathExpressionException;
-import java.io.Reader;
-import java.io.IOException;
import java.util.*;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
@@ -292,50 +290,38 @@ public final class IndexSchema {
queryAnalyzer = new SolrQueryAnalyzer();
}
- private class SolrIndexAnalyzer extends Analyzer {
- protected final HashMap<String,Analyzer> analyzers;
+ private class SolrIndexAnalyzer extends AnalyzerWrapper {
+ protected final HashMap<String, Analyzer> analyzers;
SolrIndexAnalyzer() {
analyzers = analyzerCache();
}
- protected HashMap<String,Analyzer> analyzerCache() {
- HashMap<String,Analyzer> cache = new HashMap<String,Analyzer>();
- for (SchemaField f : getFields().values()) {
+ protected HashMap<String, Analyzer> analyzerCache() {
+ HashMap<String, Analyzer> cache = new HashMap<String, Analyzer>();
+ for (SchemaField f : getFields().values()) {
Analyzer analyzer = f.getType().getAnalyzer();
cache.put(f.getName(), analyzer);
}
return cache;
}
- protected Analyzer getAnalyzer(String fieldName)
- {
- Analyzer analyzer = analyzers.get(fieldName);
- return analyzer!=null ? analyzer : getDynamicFieldType(fieldName).getAnalyzer();
- }
-
- @Override
- public TokenStream tokenStream(String fieldName, Reader reader)
- {
- return getAnalyzer(fieldName).tokenStream(fieldName,reader);
- }
-
@Override
- public TokenStream reusableTokenStream(String fieldName, Reader reader) throws IOException {
- return getAnalyzer(fieldName).reusableTokenStream(fieldName,reader);
+ protected Analyzer getWrappedAnalyzer(String fieldName) {
+ Analyzer analyzer = analyzers.get(fieldName);
+ return analyzer != null ? analyzer : getDynamicFieldType(fieldName).getAnalyzer();
}
@Override
- public int getPositionIncrementGap(String fieldName) {
- return getAnalyzer(fieldName).getPositionIncrementGap(fieldName);
+ protected TokenStreamComponents wrapComponents(String fieldName, TokenStreamComponents components) {
+ return components;
}
}
-
private class SolrQueryAnalyzer extends SolrIndexAnalyzer {
@Override
- protected HashMap<String,Analyzer> analyzerCache() {
- HashMap<String,Analyzer> cache = new HashMap<String,Analyzer>();
+ protected HashMap<String, Analyzer> analyzerCache() {
+ HashMap<String, Analyzer> cache = new HashMap<String, Analyzer>();
for (SchemaField f : getFields().values()) {
Analyzer analyzer = f.getType().getQueryAnalyzer();
cache.put(f.getName(), analyzer);
@@ -344,10 +330,9 @@ public final class IndexSchema {
}
@Override
- protected Analyzer getAnalyzer(String fieldName)
- {
+ protected Analyzer getWrappedAnalyzer(String fieldName) {
Analyzer analyzer = analyzers.get(fieldName);
- return analyzer!=null ? analyzer : getDynamicFieldType(fieldName).getQueryAnalyzer();
+ return analyzer != null ? analyzer : getDynamicFieldType(fieldName).getQueryAnalyzer();
}
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/LatLonType.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/LatLonType.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/LatLonType.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/LatLonType.java Sat Oct 1 03:04:53 2011
@@ -366,13 +366,14 @@ class SpatialDistanceQuery extends Query
}
@Override
- public Scorer scorer(AtomicReaderContext context, ScorerContext scorerContext) throws IOException {
- return new SpatialScorer(context, this, queryWeight);
+ public Scorer scorer(AtomicReaderContext context, boolean scoreDocsInOrder,
+ boolean topScorer, Bits acceptDocs) throws IOException {
+ return new SpatialScorer(context, acceptDocs, this, queryWeight);
}
@Override
public Explanation explain(AtomicReaderContext context, int doc) throws IOException {
- return ((SpatialScorer)scorer(context, ScorerContext.def().scoreDocsInOrder(true).topScorer(true))).explain(doc);
+ return ((SpatialScorer)scorer(context, true, true, context.reader.getLiveDocs())).explain(doc);
}
}
@@ -400,13 +401,13 @@ class SpatialDistanceQuery extends Query
int lastDistDoc;
double lastDist;
- public SpatialScorer(AtomicReaderContext readerContext, SpatialWeight w, float qWeight) throws IOException {
+ public SpatialScorer(AtomicReaderContext readerContext, Bits acceptDocs, SpatialWeight w, float qWeight) throws IOException {
super(w);
this.weight = w;
this.qWeight = qWeight;
this.reader = readerContext.reader;
this.maxDoc = reader.maxDoc();
- this.liveDocs = reader.getLiveDocs();
+ this.liveDocs = acceptDocs;
latVals = latSource.getValues(weight.latContext, readerContext);
lonVals = lonSource.getValues(weight.lonContext, readerContext);
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/TextField.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/TextField.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/TextField.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/schema/TextField.java Sat Oct 1 03:04:53 2011
@@ -109,30 +109,27 @@ public class TextField extends FieldType
TokenStream source;
try {
- source = analyzer.reusableTokenStream(field, new StringReader(queryText));
+ source = analyzer.tokenStream(field, new StringReader(queryText));
source.reset();
} catch (IOException e) {
- source = analyzer.tokenStream(field, new StringReader(queryText));
+ throw new RuntimeException("Unable to initialize TokenStream to analyze query text", e);
}
CachingTokenFilter buffer = new CachingTokenFilter(source);
CharTermAttribute termAtt = null;
PositionIncrementAttribute posIncrAtt = null;
int numTokens = 0;
- boolean success = false;
try {
buffer.reset();
- success = true;
} catch (IOException e) {
- // success==false if we hit an exception
+ throw new RuntimeException("Unable to initialize TokenStream to analyze query text", e);
}
- if (success) {
- if (buffer.hasAttribute(CharTermAttribute.class)) {
- termAtt = buffer.getAttribute(CharTermAttribute.class);
- }
- if (buffer.hasAttribute(PositionIncrementAttribute.class)) {
- posIncrAtt = buffer.getAttribute(PositionIncrementAttribute.class);
- }
+
+ if (buffer.hasAttribute(CharTermAttribute.class)) {
+ termAtt = buffer.getAttribute(CharTermAttribute.class);
+ }
+ if (buffer.hasAttribute(PositionIncrementAttribute.class)) {
+ posIncrAtt = buffer.getAttribute(PositionIncrementAttribute.class);
}
int positionCount = 0;
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/Grouping.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/Grouping.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/Grouping.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/Grouping.java Sat Oct 1 03:04:53 2011
@@ -666,7 +666,7 @@ public class Grouping {
}
sort = sort == null ? Sort.RELEVANCE : sort;
- firstPass = new TermFirstPassGroupingCollectorJava6(groupBy, sort, actualGroupsToFind);
+ firstPass = new TermFirstPassGroupingCollector(groupBy, sort, actualGroupsToFind);
return firstPass;
}
@@ -1013,22 +1013,6 @@ public class Grouping {
filler = docValues.getValueFiller();
mval = filler.getValue();
}
-
- @Override
- protected CollectedSearchGroup<MutableValue> pollLast() {
- return orderedGroups.pollLast();
- }
- }
-
- static class TermFirstPassGroupingCollectorJava6 extends TermFirstPassGroupingCollector {
- public TermFirstPassGroupingCollectorJava6(String groupField, Sort groupSort, int topNGroups) throws IOException {
- super(groupField, groupSort, topNGroups);
- }
-
- @Override
- protected CollectedSearchGroup<BytesRef> pollLast() {
- return orderedGroups.pollLast();
- }
}
static class FunctionSecondPassGroupingCollector extends AbstractSecondPassGroupingCollector<MutableValue> {
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java Sat Oct 1 03:04:53 2011
@@ -187,7 +187,8 @@ class JoinQuery extends Query {
@Override
- public Scorer scorer(IndexReader.AtomicReaderContext context, ScorerContext scorerContext) throws IOException {
+ public Scorer scorer(IndexReader.AtomicReaderContext context, boolean scoreDocsInOrder,
+ boolean topScorer, Bits acceptDocs) throws IOException {
if (filter == null) {
boolean debug = rb != null && rb.isDebug();
long start = debug ? System.currentTimeMillis() : 0;
@@ -483,7 +484,7 @@ class JoinQuery extends Query {
@Override
public Explanation explain(IndexReader.AtomicReaderContext context, int doc) throws IOException {
- Scorer scorer = scorer(context, null);
+ Scorer scorer = scorer(context, true, false, context.reader.getLiveDocs());
boolean exists = scorer.advance(doc) == doc;
ComplexExplanation result = new ComplexExplanation();
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrConstantScoreQuery.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrConstantScoreQuery.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrConstantScoreQuery.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrConstantScoreQuery.java Sat Oct 1 03:04:53 2011
@@ -2,6 +2,7 @@ package org.apache.solr.search;
import org.apache.lucene.queries.function.ValueSource;
import org.apache.lucene.search.*;
+import org.apache.lucene.util.Bits;
import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.IndexReader.AtomicReaderContext;
import org.apache.solr.common.SolrException;
@@ -118,7 +119,8 @@ public class SolrConstantScoreQuery exte
}
@Override
- public Scorer scorer(AtomicReaderContext context, ScorerContext scorerContext) throws IOException {
+ public Scorer scorer(AtomicReaderContext context, boolean scoreDocsInOrder,
+ boolean topScorer, Bits acceptDocs) throws IOException {
return new ConstantScorer(context, this, queryWeight);
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java Sat Oct 1 03:04:53 2011
@@ -2080,7 +2080,7 @@ class FilterImpl extends Filter {
iterators.add(iter);
}
for (Weight w : weights) {
- Scorer scorer = w.scorer(context, Weight.ScorerContext.def());
+ Scorer scorer = w.scorer(context, true, false, context.reader.getLiveDocs());
if (scorer == null) return null;
iterators.add(scorer);
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/ValueSourceParser.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/ValueSourceParser.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/ValueSourceParser.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/search/ValueSourceParser.java Sat Oct 1 03:04:53 2011
@@ -483,19 +483,20 @@ public abstract class ValueSourceParser
return Math.atan2(a.doubleVal(doc), b.doubleVal(doc));
}
});
- addParser(new Double2Parser("max") {
+ addParser("max", new ValueSourceParser() {
@Override
- public double func(int doc, DocValues a, DocValues b) {
- return Math.max(a.doubleVal(doc), b.doubleVal(doc));
+ public ValueSource parse(FunctionQParser fp) throws ParseException {
+ List<ValueSource> sources = fp.parseValueSourceList();
+ return new MaxFloatFunction(sources.toArray(new ValueSource[sources.size()]));
}
});
- addParser(new Double2Parser("min") {
+ addParser("min", new ValueSourceParser() {
@Override
- public double func(int doc, DocValues a, DocValues b) {
- return Math.min(a.doubleVal(doc), b.doubleVal(doc));
+ public ValueSource parse(FunctionQParser fp) throws ParseException {
+ List<ValueSource> sources = fp.parseValueSourceList();
+ return new MinFloatFunction(sources.toArray(new ValueSource[sources.size()]));
}
});
-
addParser("sqedist", new ValueSourceParser() {
@Override
public ValueSource parse(FunctionQParser fp) throws ParseException {
Modified: lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/spelling/SpellingQueryConverter.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/spelling/SpellingQueryConverter.java?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/spelling/SpellingQueryConverter.java (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/java/org/apache/solr/spelling/SpellingQueryConverter.java Sat Oct 1 03:04:53 2011
@@ -104,7 +104,7 @@ public class SpellingQueryConverter exte
String word = matcher.group(0);
if (word.equals("AND") == false && word.equals("OR") == false) {
try {
- stream = analyzer.reusableTokenStream("", new StringReader(word));
+ stream = analyzer.tokenStream("", new StringReader(word));
// TODO: support custom attributes
CharTermAttribute termAtt = stream.addAttribute(CharTermAttribute.class);
FlagsAttribute flagsAtt = stream.addAttribute(FlagsAttribute.class);
@@ -123,6 +123,8 @@ public class SpellingQueryConverter exte
token.setPositionIncrement(posIncAtt.getPositionIncrement());
result.add(token);
}
+ stream.end();
+ stream.close();
} catch (IOException e) {
}
}
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-copyfield-test.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-copyfield-test.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-copyfield-test.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-copyfield-test.xml Sat Oct 1 03:04:53 2011
@@ -76,7 +76,7 @@
<!-- Field type demonstrating an Analyzer failure -->
<fieldtype name="failtype1" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="0" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -85,7 +85,7 @@
<!-- Demonstrating ignoreCaseChange -->
<fieldtype name="wdf_nocase" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="0" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -135,7 +135,7 @@
<analyzer><tokenizer class="solr.LowerCaseTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="keywordtok" class="solr.TextField">
- <analyzer><tokenizer class="solr.KeywordTokenizerFactory"/></analyzer>
+ <analyzer><tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/></analyzer>
</fieldtype>
<fieldtype name="standardtok" class="solr.TextField">
<analyzer><tokenizer class="solr.StandardTokenizerFactory"/></analyzer>
@@ -144,7 +144,7 @@
<analyzer><tokenizer class="solr.LetterTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="whitetok" class="solr.TextField">
- <analyzer><tokenizer class="solr.WhitespaceTokenizerFactory"/></analyzer>
+ <analyzer><tokenizer class="solr.MockTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="HTMLstandardtok" class="solr.TextField">
<analyzer>
@@ -155,7 +155,7 @@
<fieldtype name="HTMLwhitetok" class="solr.TextField">
<analyzer>
<charFilter class="solr.HTMLStripCharFilterFactory"/>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="standardtokfilt" class="solr.TextField">
@@ -166,81 +166,81 @@
</fieldtype>
<fieldtype name="standardfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StandardFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="lowerfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="patternreplacefilt" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
<filter class="solr.PatternReplaceFilterFactory"
pattern="([^a-zA-Z])" replacement="_" replace="all"
/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
</analyzer>
</fieldtype>
<fieldtype name="porterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<!-- fieldtype name="snowballfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SnowballPorterFilterFactory"/>
</analyzer>
</fieldtype -->
<fieldtype name="engporterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="custengporterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.KeywordMarkerFilterFactory" protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="stopfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true"/>
</analyzer>
</fieldtype>
<fieldtype name="custstopfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StopFilterFactory" words="stopwords.txt"/>
</analyzer>
</fieldtype>
<fieldtype name="lengthfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LengthFilterFactory" min="2" max="5"/>
</analyzer>
</fieldtype>
<fieldtype name="subword" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory"/>
@@ -251,12 +251,12 @@
<!-- more flexible in matching skus, but more chance of a false match -->
<fieldtype name="skutype1" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -265,12 +265,12 @@
<!-- less flexible in matching skus, but less chance of a false match -->
<fieldtype name="skutype2" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -279,7 +279,7 @@
<!-- less flexible in matching skus, but less chance of a false match -->
<fieldtype name="syn" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter name="syn" class="solr.SynonymFilterFactory" synonyms="synonyms.txt"/>
</analyzer>
</fieldtype>
@@ -289,7 +289,7 @@
-->
<fieldtype name="dedup" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory"
synonyms="synonyms.txt" expand="true" />
<filter class="solr.PorterStemFilterFactory"/>
@@ -302,7 +302,7 @@
<fieldtype name="textgap" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-required-fields.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-required-fields.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-required-fields.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-required-fields.xml Sat Oct 1 03:04:53 2011
@@ -76,7 +76,7 @@
<!-- Field type demonstrating an Analyzer failure -->
<fieldtype name="failtype1" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="0" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -126,7 +126,7 @@
<analyzer><tokenizer class="solr.LowerCaseTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="keywordtok" class="solr.TextField">
- <analyzer><tokenizer class="solr.KeywordTokenizerFactory"/></analyzer>
+ <analyzer><tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/></analyzer>
</fieldtype>
<fieldtype name="standardtok" class="solr.TextField">
<analyzer><tokenizer class="solr.StandardTokenizerFactory"/></analyzer>
@@ -135,7 +135,7 @@
<analyzer><tokenizer class="solr.LetterTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="whitetok" class="solr.TextField">
- <analyzer><tokenizer class="solr.WhitespaceTokenizerFactory"/></analyzer>
+ <analyzer><tokenizer class="solr.MockTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="HTMLstandardtok" class="solr.TextField">
<analyzer>
@@ -146,7 +146,7 @@
<fieldtype name="HTMLwhitetok" class="solr.TextField">
<analyzer>
<charFilter class="solr.HTMLStripCharFilterFactory"/>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="standardtokfilt" class="solr.TextField">
@@ -157,81 +157,81 @@
</fieldtype>
<fieldtype name="standardfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StandardFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="lowerfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="patternreplacefilt" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
<filter class="solr.PatternReplaceFilterFactory"
pattern="([^a-zA-Z])" replacement="_" replace="all"
/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
</analyzer>
</fieldtype>
<fieldtype name="porterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<!-- fieldtype name="snowballfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SnowballPorterFilterFactory"/>
</analyzer>
</fieldtype -->
<fieldtype name="engporterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="custengporterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.KeywordMarkerFilterFactory" protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="stopfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true"/>
</analyzer>
</fieldtype>
<fieldtype name="custstopfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StopFilterFactory" words="stopwords.txt"/>
</analyzer>
</fieldtype>
<fieldtype name="lengthfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LengthFilterFactory" min="2" max="5"/>
</analyzer>
</fieldtype>
<fieldtype name="subword" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory"/>
@@ -242,12 +242,12 @@
<!-- more flexible in matching skus, but more chance of a false match -->
<fieldtype name="skutype1" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -256,12 +256,12 @@
<!-- less flexible in matching skus, but less chance of a false match -->
<fieldtype name="skutype2" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -270,7 +270,7 @@
<!-- less flexible in matching skus, but less chance of a false match -->
<fieldtype name="syn" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter name="syn" class="solr.SynonymFilterFactory" synonyms="synonyms.txt"/>
</analyzer>
</fieldtype>
@@ -280,7 +280,7 @@
-->
<fieldtype name="dedup" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory"
synonyms="synonyms.txt" expand="true" />
<filter class="solr.PorterStemFilterFactory"/>
@@ -293,7 +293,7 @@
<fieldtype name="textgap" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-reversed.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-reversed.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-reversed.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-reversed.xml Sat Oct 1 03:04:53 2011
@@ -30,28 +30,28 @@
<fieldtype name="text" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="srev" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.ReversedWildcardFilterFactory" withOriginal="true"
maxPosAsterisk="3" maxPosQuestion="2" maxFractionAsterisk="0.33"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="rev" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.ReversedWildcardFilterFactory" withOriginal="false"
maxPosAsterisk="3" maxPosQuestion="2" maxFractionAsterisk="0.33"
@@ -59,7 +59,7 @@
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-stop-keep.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-stop-keep.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-stop-keep.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-stop-keep.xml Sat Oct 1 03:04:53 2011
@@ -34,7 +34,7 @@
<fieldtype name="string" class="solr.StrField" sortMissingLast="true"/>
<fieldtype name="stop-one" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true"
words="stop-1.txt"/>
@@ -43,7 +43,7 @@
</fieldtype>
<fieldtype name="stop-two" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true"
words="stop-1.txt,stop-2.txt"/>
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-trie.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-trie.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-trie.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema-trie.xml Sat Oct 1 03:04:53 2011
@@ -156,7 +156,7 @@
<!-- A text field that only splits on whitespace for exact matching of words -->
<fieldType name="text_ws" class="solr.TextField" positionIncrementGap="100">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldType>
@@ -169,7 +169,7 @@
-->
<fieldType name="text" class="solr.TextField" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<!-- in this example, we will only use synonyms at query time
<filter class="solr.SynonymFilterFactory" synonyms="index_synonyms.txt" ignoreCase="true" expand="false"/>
-->
@@ -181,7 +181,7 @@
<filter class="solr.RemoveDuplicatesTokenFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="true"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="1"/>
@@ -197,7 +197,7 @@
but may be good for SKUs. Can insert dashes in the wrong place and still match. -->
<fieldType name="textTight" class="solr.TextField" positionIncrementGap="100" >
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="false"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
@@ -217,7 +217,7 @@
<!-- KeywordTokenizer does no actual tokenizing, so the entire
input string is preserved as a single token
-->
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
<!-- The LowerCase TokenFilter does what you expect, which can be
when you want your sorting to be case insensitive
-->
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema.xml Sat Oct 1 03:04:53 2011
@@ -86,7 +86,7 @@
<!-- Field type demonstrating an Analyzer failure -->
<fieldtype name="failtype1" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="0" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -95,7 +95,7 @@
<!-- Demonstrating ignoreCaseChange -->
<fieldtype name="wdf_nocase" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="0" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="0" preserveOriginal="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -103,7 +103,7 @@
<fieldtype name="wdf_preserve" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="0" preserveOriginal="1"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -156,7 +156,7 @@
<analyzer><tokenizer class="solr.LowerCaseTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="keywordtok" class="solr.TextField">
- <analyzer><tokenizer class="solr.KeywordTokenizerFactory"/></analyzer>
+ <analyzer><tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/></analyzer>
</fieldtype>
<fieldtype name="standardtok" class="solr.TextField">
<analyzer><tokenizer class="solr.StandardTokenizerFactory"/></analyzer>
@@ -165,7 +165,7 @@
<analyzer><tokenizer class="solr.LetterTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="whitetok" class="solr.TextField">
- <analyzer><tokenizer class="solr.WhitespaceTokenizerFactory"/></analyzer>
+ <analyzer><tokenizer class="solr.MockTokenizerFactory"/></analyzer>
</fieldtype>
<fieldtype name="HTMLstandardtok" class="solr.TextField">
<analyzer>
@@ -176,7 +176,7 @@
<fieldtype name="HTMLwhitetok" class="solr.TextField">
<analyzer>
<charFilter class="solr.HTMLStripCharFilterFactory"/>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="standardtokfilt" class="solr.TextField">
@@ -187,19 +187,19 @@
</fieldtype>
<fieldtype name="standardfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StandardFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="lowerfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="lowerpunctfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter name="syn" class="solr.SynonymFilterFactory" synonyms="synonyms.txt" expand="true"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="1" splitOnCaseChange="1"/>
<filter class="solr.LowerCaseFilterFactory"/>
@@ -207,13 +207,13 @@
</fieldtype>
<fieldtype name="patternreplacefilt" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
<filter class="solr.PatternReplaceFilterFactory"
pattern="([^a-zA-Z])" replacement="_" replace="all"
/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
</analyzer>
</fieldtype>
<fieldtype name="patterntok" class="solr.TextField">
@@ -223,44 +223,44 @@
</fieldtype>
<fieldtype name="porterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<!-- fieldtype name="snowballfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SnowballPorterFilterFactory"/>
</analyzer>
</fieldtype -->
<fieldtype name="engporterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="custengporterfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.KeywordMarkerFilterFactory" protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldtype>
<fieldtype name="stopfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true"/>
</analyzer>
</fieldtype>
<fieldtype name="custstopfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.StopFilterFactory" words="stopwords.txt"/>
</analyzer>
</fieldtype>
<fieldtype name="lengthfilt" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LengthFilterFactory" min="2" max="5"/>
</analyzer>
</fieldtype>
@@ -268,20 +268,20 @@
<analyzer>
<charFilter class="solr.HTMLStripCharFilterFactory"/>
<charFilter class="solr.MappingCharFilterFactory" mapping="mapping-ISOLatin1Accent.txt"/>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldType>
<fieldtype name="subword" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.StopFilterFactory"/>
@@ -291,14 +291,14 @@
<fieldtype name="numericsubword" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.WordDelimiterFilterFactory" protected="protwords.txt" splitOnNumerics="0" splitOnCaseChange="0" generateWordParts="1" generateNumberParts="0" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
<filter class="solr.StopFilterFactory"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.WordDelimiterFilterFactory" protected="protwords.txt" splitOnNumerics="0" splitOnCaseChange="0" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.StopFilterFactory"/>
@@ -308,12 +308,12 @@
<fieldtype name="protectedsubword" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.WordDelimiterFilterFactory" protected="protwords.txt" splitOnNumerics="0" splitOnCaseChange="0" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
@@ -322,12 +322,12 @@
<!-- more flexible in matching skus, but more chance of a false match -->
<fieldtype name="skutype1" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -336,12 +336,12 @@
<!-- less flexible in matching skus, but less chance of a false match -->
<fieldtype name="skutype2" class="solr.TextField">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
@@ -350,7 +350,7 @@
<!-- less flexible in matching skus, but less chance of a false match -->
<fieldtype name="syn" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter name="syn" class="solr.SynonymFilterFactory" synonyms="old_synonyms.txt"/>
</analyzer>
</fieldtype>
@@ -360,7 +360,7 @@
-->
<fieldtype name="dedup" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory"
synonyms="old_synonyms.txt" expand="true" />
<filter class="solr.PorterStemFilterFactory"/>
@@ -373,7 +373,7 @@
<fieldtype name="textgap" class="solr.TextField" multiValued="true" positionIncrementGap="100">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldtype>
@@ -396,7 +396,7 @@
<!-- specify a Similarity classname directly -->
<fieldType name="sim1" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
<similarity class="org.apache.lucene.misc.SweetSpotSimilarity"/>
</fieldType>
@@ -404,7 +404,7 @@
<!-- specify a Similarity factory -->
<fieldType name="sim2" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
<similarity class="solr.CustomSimilarityFactory">
<str name="echo">is there an echo?</str>
@@ -414,14 +414,14 @@
<!-- don't specify any sim at all: get the default -->
<fieldType name="sim3" class="solr.TextField">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldType>
<!-- omitPositions example -->
<fieldType name="nopositions" class="solr.TextField" omitPositions="true">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldType>
</types>
Modified: lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema11.xml
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema11.xml?rev=1177888&r1=1177887&r2=1177888&view=diff
==============================================================================
--- lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema11.xml (original)
+++ lucene/dev/branches/lucene2621/solr/core/src/test-files/solr/conf/schema11.xml Sat Oct 1 03:04:53 2011
@@ -147,7 +147,7 @@
<!-- A text field that only splits on whitespace for exact matching of words -->
<fieldType name="text_ws" class="solr.TextField" positionIncrementGap="100">
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
</analyzer>
</fieldType>
@@ -160,7 +160,7 @@
-->
<fieldType name="text" class="solr.TextField" positionIncrementGap="100">
<analyzer type="index">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<!-- in this example, we will only use synonyms at query time
<filter class="solr.SynonymFilterFactory" synonyms="index_synonyms.txt" ignoreCase="true" expand="false"/>
-->
@@ -172,7 +172,7 @@
<filter class="solr.RemoveDuplicatesTokenFilterFactory"/>
</analyzer>
<analyzer type="query">
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="true"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="1"/>
@@ -188,7 +188,7 @@
but may be good for SKUs. Can insert dashes in the wrong place and still match. -->
<fieldType name="textTight" class="solr.TextField" positionIncrementGap="100" >
<analyzer>
- <tokenizer class="solr.WhitespaceTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="false"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" generateNumberParts="0" catenateWords="1" catenateNumbers="1" catenateAll="0"/>
@@ -208,7 +208,7 @@
<!-- KeywordTokenizer does no actual tokenizing, so the entire
input string is preserved as a single token
-->
- <tokenizer class="solr.KeywordTokenizerFactory"/>
+ <tokenizer class="solr.MockTokenizerFactory" pattern="keyword"/>
<!-- The LowerCase TokenFilter does what you expect, which can be
when you want your sorting to be case insensitive
-->