You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by no...@apache.org on 2016/09/07 09:56:54 UTC
[06/50] [abbrv] lucene-solr:apiv2: simplify test to use
CannedTokenStream
simplify test to use CannedTokenStream
Project: http://git-wip-us.apache.org/repos/asf/lucene-solr/repo
Commit: http://git-wip-us.apache.org/repos/asf/lucene-solr/commit/13acba8b
Tree: http://git-wip-us.apache.org/repos/asf/lucene-solr/tree/13acba8b
Diff: http://git-wip-us.apache.org/repos/asf/lucene-solr/diff/13acba8b
Branch: refs/heads/apiv2
Commit: 13acba8b4e712222a35f61412ed5b1f45cbec3ee
Parents: f6253d5
Author: Robert Muir <rm...@apache.org>
Authored: Thu Aug 25 12:07:53 2016 -0400
Committer: Robert Muir <rm...@apache.org>
Committed: Thu Aug 25 12:07:53 2016 -0400
----------------------------------------------------------------------
.../analysis/miscellaneous/TestTrimFilter.java | 46 +-------------------
1 file changed, 2 insertions(+), 44 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/13acba8b/lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestTrimFilter.java
----------------------------------------------------------------------
diff --git a/lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestTrimFilter.java b/lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestTrimFilter.java
index 75a6e74..380f52a 100644
--- a/lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestTrimFilter.java
+++ b/lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestTrimFilter.java
@@ -20,17 +20,12 @@ import java.io.IOException;
import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.BaseTokenStreamTestCase;
+import org.apache.lucene.analysis.CannedTokenStream;
import org.apache.lucene.analysis.MockTokenizer;
import org.apache.lucene.analysis.Token;
import org.apache.lucene.analysis.TokenStream;
import org.apache.lucene.analysis.Tokenizer;
import org.apache.lucene.analysis.core.KeywordTokenizer;
-import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
-import org.apache.lucene.analysis.tokenattributes.FlagsAttribute;
-import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
-import org.apache.lucene.analysis.tokenattributes.PayloadAttribute;
-import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute;
-import org.apache.lucene.analysis.tokenattributes.TypeAttribute;
/**
*/
@@ -43,7 +38,7 @@ public class TestTrimFilter extends BaseTokenStreamTestCase {
char[] whitespace = " ".toCharArray();
char[] empty = "".toCharArray();
- TokenStream ts = new IterTokenStream(new Token(new String(a, 0, a.length), 1, 5),
+ TokenStream ts = new CannedTokenStream(new Token(new String(a, 0, a.length), 1, 5),
new Token(new String(b, 0, b.length), 6, 10),
new Token(new String(ccc, 0, ccc.length), 11, 15),
new Token(new String(whitespace, 0, whitespace.length), 16, 20),
@@ -53,43 +48,6 @@ public class TestTrimFilter extends BaseTokenStreamTestCase {
assertTokenStreamContents(ts, new String[] { "a", "b", "cCc", "", ""});
}
- /**
- * @deprecated (3.0) does not support custom attributes
- */
- @Deprecated
- private static class IterTokenStream extends TokenStream {
- final Token tokens[];
- int index = 0;
- CharTermAttribute termAtt = addAttribute(CharTermAttribute.class);
- OffsetAttribute offsetAtt = addAttribute(OffsetAttribute.class);
- PositionIncrementAttribute posIncAtt = addAttribute(PositionIncrementAttribute.class);
- FlagsAttribute flagsAtt = addAttribute(FlagsAttribute.class);
- TypeAttribute typeAtt = addAttribute(TypeAttribute.class);
- PayloadAttribute payloadAtt = addAttribute(PayloadAttribute.class);
-
- public IterTokenStream(Token... tokens) {
- super();
- this.tokens = tokens;
- }
-
- @Override
- public boolean incrementToken() throws IOException {
- if (index >= tokens.length)
- return false;
- else {
- clearAttributes();
- Token token = tokens[index++];
- termAtt.setEmpty().append(token);
- offsetAtt.setOffset(token.startOffset(), token.endOffset());
- posIncAtt.setPositionIncrement(token.getPositionIncrement());
- flagsAtt.setFlags(token.getFlags());
- typeAtt.setType(token.type());
- payloadAtt.setPayload(token.getPayload());
- return true;
- }
- }
- }
-
/** blast some random strings through the analyzer */
public void testRandomStrings() throws Exception {
Analyzer a = new Analyzer() {