You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucenenet.apache.org by ni...@apache.org on 2017/02/26 23:37:38 UTC
[50/72] [abbrv] [partial] lucenenet git commit: Lucene.Net.Tests:
Removed \core directory and put its contents in root directory
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestGraphTokenizers.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestGraphTokenizers.cs b/src/Lucene.Net.Tests/Analysis/TestGraphTokenizers.cs
new file mode 100644
index 0000000..73619d1
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestGraphTokenizers.cs
@@ -0,0 +1,728 @@
+using Lucene.Net.Analysis.TokenAttributes;
+using System;
+using System.Collections.Generic;
+using System.Diagnostics;
+using System.Text;
+
+namespace Lucene.Net.Analysis
+{
+ using Lucene.Net.Support;
+ using NUnit.Framework;
+ using System.IO;
+
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ using Automaton = Lucene.Net.Util.Automaton.Automaton;
+ using BasicAutomata = Lucene.Net.Util.Automaton.BasicAutomata;
+ using BasicOperations = Lucene.Net.Util.Automaton.BasicOperations;
+
+ [TestFixture]
+ public class TestGraphTokenizers : BaseTokenStreamTestCase
+ {
+ // Makes a graph TokenStream from the string; separate
+ // positions with single space, multiple tokens at the same
+ // position with /, and add optional position length with
+ // :. EG "a b c" is a simple chain, "a/x b c" adds 'x'
+ // over 'a' at position 0 with posLen=1, "a/x:3 b c" adds
+ // 'x' over a with posLen=3. Tokens are in normal-form!
+ // So, offsets are computed based on the first token at a
+ // given position. NOTE: each token must be a single
+ // character! We assume this when computing offsets...
+
+ // NOTE: all input tokens must be length 1!!! this means
+ // you cannot turn on MockCharFilter when random
+ // testing...
+
+ private class GraphTokenizer : Tokenizer
+ {
+ internal IList<Token> Tokens;
+ internal int Upto;
+ internal int InputLength;
+
+ internal readonly ICharTermAttribute TermAtt;
+ internal readonly IOffsetAttribute OffsetAtt;
+ internal readonly IPositionIncrementAttribute PosIncrAtt;
+ internal readonly IPositionLengthAttribute PosLengthAtt;
+
+ public GraphTokenizer(TextReader input)
+ : base(input)
+ {
+ TermAtt = AddAttribute<ICharTermAttribute>();
+ OffsetAtt = AddAttribute<IOffsetAttribute>();
+ PosIncrAtt = AddAttribute<IPositionIncrementAttribute>();
+ PosLengthAtt = AddAttribute<IPositionLengthAttribute>();
+ }
+
+ public override void Reset()
+ {
+ base.Reset();
+ Tokens = null;
+ Upto = 0;
+ }
+
+ public sealed override bool IncrementToken()
+ {
+ if (Tokens == null)
+ {
+ FillTokens();
+ }
+ //System.out.println("graphTokenizer: incr upto=" + upto + " vs " + tokens.size());
+ if (Upto == Tokens.Count)
+ {
+ //System.out.println(" END @ " + tokens.size());
+ return false;
+ }
+ Token t = Tokens[Upto++];
+ //System.out.println(" return token=" + t);
+ ClearAttributes();
+ TermAtt.Append(t.ToString());
+ OffsetAtt.SetOffset(t.StartOffset, t.EndOffset);
+ PosIncrAtt.PositionIncrement = t.PositionIncrement;
+ PosLengthAtt.PositionLength = t.PositionLength;
+ return true;
+ }
+
+ public override void End()
+ {
+ base.End();
+ // NOTE: somewhat... hackish, but we need this to
+ // satisfy BTSTC:
+ int lastOffset;
+ if (Tokens != null && Tokens.Count > 0)
+ {
+ lastOffset = Tokens[Tokens.Count - 1].EndOffset;
+ }
+ else
+ {
+ lastOffset = 0;
+ }
+ OffsetAtt.SetOffset(CorrectOffset(lastOffset), CorrectOffset(InputLength));
+ }
+
+ internal virtual void FillTokens()
+ {
+ StringBuilder sb = new StringBuilder();
+ char[] buffer = new char[256];
+ while (true)
+ {
+ int count = m_input.Read(buffer, 0, buffer.Length);
+
+ //.NET TextReader.Read(buff, int, int) returns 0, not -1 on no chars
+ // but in some cases, such as MockCharFilter, it overloads read and returns -1
+ // so we should handle both 0 and -1 values
+ if (count <= 0)
+ {
+ break;
+ }
+ sb.Append(buffer, 0, count);
+ //System.out.println("got count=" + count);
+ }
+ //System.out.println("fillTokens: " + sb);
+
+ InputLength = sb.Length;
+
+ string[] parts = sb.ToString().Split(' ');
+
+ Tokens = new List<Token>();
+ int pos = 0;
+ int maxPos = -1;
+ int offset = 0;
+ //System.out.println("again");
+ foreach (string part in parts)
+ {
+ string[] overlapped = part.Split('/');
+ bool firstAtPos = true;
+ int minPosLength = int.MaxValue;
+ foreach (string part2 in overlapped)
+ {
+ int colonIndex = part2.IndexOf(':');
+ string token;
+ int posLength;
+ if (colonIndex != -1)
+ {
+ token = part2.Substring(0, colonIndex);
+ posLength = Convert.ToInt32(part2.Substring(1 + colonIndex));
+ }
+ else
+ {
+ token = part2;
+ posLength = 1;
+ }
+ maxPos = Math.Max(maxPos, pos + posLength);
+ minPosLength = Math.Min(minPosLength, posLength);
+ Token t = new Token(token, offset, offset + 2 * posLength - 1);
+ t.PositionLength = posLength;
+ t.PositionIncrement = firstAtPos ? 1 : 0;
+ firstAtPos = false;
+ //System.out.println(" add token=" + t + " startOff=" + t.StartOffset + " endOff=" + t.EndOffset);
+ Tokens.Add(t);
+ }
+ pos += minPosLength;
+ offset = 2 * pos;
+ }
+ Debug.Assert(maxPos <= pos, "input string mal-formed: posLength>1 tokens hang over the end");
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterBasic()
+ {
+ for (int iter = 0; iter < 10 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper(this);
+
+ CheckAnalysisConsistency(Random(), a, false, "a b c d e f g h i j k");
+ }
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t2 = new MockGraphTokenFilter(Random(), t);
+ return new TokenStreamComponents(t, t2);
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterOnGraphInput()
+ {
+ for (int iter = 0; iter < 100 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper2(this);
+
+ CheckAnalysisConsistency(Random(), a, false, "a/x:3 c/y:2 d e f/z:4 g h i j k");
+ }
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper2 : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper2(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new GraphTokenizer(reader);
+ TokenStream t2 = new MockGraphTokenFilter(Random(), t);
+ return new TokenStreamComponents(t, t2);
+ }
+ }
+
+ // Just deletes (leaving hole) token 'a':
+ private sealed class RemoveATokens : TokenFilter
+ {
+ internal int PendingPosInc;
+
+ internal readonly ICharTermAttribute TermAtt;// = addAttribute(typeof(CharTermAttribute));
+ internal readonly IPositionIncrementAttribute PosIncAtt;// = addAttribute(typeof(PositionIncrementAttribute));
+
+ public RemoveATokens(TokenStream @in)
+ : base(@in)
+ {
+ TermAtt = AddAttribute<ICharTermAttribute>();
+ PosIncAtt = AddAttribute<IPositionIncrementAttribute>();
+ }
+
+ public override void Reset()
+ {
+ base.Reset();
+ PendingPosInc = 0;
+ }
+
+ public override void End()
+ {
+ base.End();
+ PosIncAtt.PositionIncrement = PendingPosInc + PosIncAtt.PositionIncrement;
+ }
+
+ public override bool IncrementToken()
+ {
+ while (true)
+ {
+ bool gotOne = m_input.IncrementToken();
+ if (!gotOne)
+ {
+ return false;
+ }
+ else if (TermAtt.ToString().Equals("a"))
+ {
+ PendingPosInc += PosIncAtt.PositionIncrement;
+ }
+ else
+ {
+ PosIncAtt.PositionIncrement = PendingPosInc + PosIncAtt.PositionIncrement;
+ PendingPosInc = 0;
+ return true;
+ }
+ }
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterBeforeHoles()
+ {
+ for (int iter = 0; iter < 100 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new MGTFBHAnalyzerAnonymousInnerClassHelper(this);
+
+ Random random = Random();
+ CheckAnalysisConsistency(random, a, false, "a b c d e f g h i j k");
+ CheckAnalysisConsistency(random, a, false, "x y a b c d e f g h i j k");
+ CheckAnalysisConsistency(random, a, false, "a b c d e f g h i j k a");
+ CheckAnalysisConsistency(random, a, false, "a b c d e f g h i j k a x y");
+ }
+ }
+
+ private class MGTFBHAnalyzerAnonymousInnerClassHelper : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public MGTFBHAnalyzerAnonymousInnerClassHelper(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t2 = new MockGraphTokenFilter(Random(), t);
+ TokenStream t3 = new RemoveATokens(t2);
+ return new TokenStreamComponents(t, t3);
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterAfterHoles()
+ {
+ for (int iter = 0; iter < 100 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new MGTFAHAnalyzerAnonymousInnerClassHelper2(this);
+
+ Random random = Random();
+ CheckAnalysisConsistency(random, a, false, "a b c d e f g h i j k");
+ CheckAnalysisConsistency(random, a, false, "x y a b c d e f g h i j k");
+ CheckAnalysisConsistency(random, a, false, "a b c d e f g h i j k a");
+ CheckAnalysisConsistency(random, a, false, "a b c d e f g h i j k a x y");
+ }
+ }
+
+ private class MGTFAHAnalyzerAnonymousInnerClassHelper2 : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public MGTFAHAnalyzerAnonymousInnerClassHelper2(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t2 = new RemoveATokens(t);
+ TokenStream t3 = new MockGraphTokenFilter(Random(), t2);
+ return new TokenStreamComponents(t, t3);
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterRandom()
+ {
+ for (int iter = 0; iter < 10 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper3(this);
+
+ Random random = Random();
+ CheckRandomData(random, a, 5, AtLeast(100));
+ }
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper3 : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper3(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t2 = new MockGraphTokenFilter(Random(), t);
+ return new TokenStreamComponents(t, t2);
+ }
+ }
+
+ // Two MockGraphTokenFilters
+ [Test]
+ public virtual void TestDoubleMockGraphTokenFilterRandom()
+ {
+ for (int iter = 0; iter < 10 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper4(this);
+
+ Random random = Random();
+ CheckRandomData(random, a, 5, AtLeast(100));
+ }
+ }
+
+ [Test]
+ public void TestMockTokenizerCtor()
+ {
+ var sr = new StringReader("Hello");
+ var mt = new MockTokenizer(sr);
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper4 : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper4(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t1 = new MockGraphTokenFilter(Random(), t);
+ TokenStream t2 = new MockGraphTokenFilter(Random(), t1);
+ return new TokenStreamComponents(t, t2);
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterBeforeHolesRandom()
+ {
+ for (int iter = 0; iter < 10 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper5(this);
+
+ Random random = Random();
+ CheckRandomData(random, a, 5, AtLeast(100));
+ }
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper5 : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper5(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t1 = new MockGraphTokenFilter(Random(), t);
+ TokenStream t2 = new MockHoleInjectingTokenFilter(Random(), t1);
+ return new TokenStreamComponents(t, t2);
+ }
+ }
+
+ [Test]
+ public virtual void TestMockGraphTokenFilterAfterHolesRandom()
+ {
+ for (int iter = 0; iter < 10 * RANDOM_MULTIPLIER; iter++)
+ {
+ if (VERBOSE)
+ {
+ Console.WriteLine("\nTEST: iter=" + iter);
+ }
+
+ // Make new analyzer each time, because MGTF has fixed
+ // seed:
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper6(this);
+
+ Random random = Random();
+ CheckRandomData(random, a, 5, AtLeast(100));
+ }
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper6 : Analyzer
+ {
+ private readonly TestGraphTokenizers OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper6(TestGraphTokenizers outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TokenStream t1 = new MockHoleInjectingTokenFilter(Random(), t);
+ TokenStream t2 = new MockGraphTokenFilter(Random(), t1);
+ return new TokenStreamComponents(t, t2);
+ }
+ }
+
+ private static Token Token(string term, int posInc, int posLength)
+ {
+ Token t = new Token(term, 0, 0);
+ t.PositionIncrement = posInc;
+ t.PositionLength = posLength;
+ return t;
+ }
+
+ private static Token Token(string term, int posInc, int posLength, int startOffset, int endOffset)
+ {
+ Token t = new Token(term, startOffset, endOffset);
+ t.PositionIncrement = posInc;
+ t.PositionLength = posLength;
+ return t;
+ }
+
+ [Test]
+ public virtual void TestSingleToken()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton expected = BasicAutomata.MakeString("abc");
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestMultipleHoles()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("a", 1, 1), Token("b", 3, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton expected = Join(S2a("a"), SEP_A, HOLE_A, SEP_A, HOLE_A, SEP_A, S2a("b"));
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestSynOverMultipleHoles()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("a", 1, 1), Token("x", 0, 3), Token("b", 3, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton a1 = Join(S2a("a"), SEP_A, HOLE_A, SEP_A, HOLE_A, SEP_A, S2a("b"));
+ Automaton a2 = Join(S2a("x"), SEP_A, S2a("b"));
+ Automaton expected = BasicOperations.Union(a1, a2);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ // for debugging!
+ /*
+ private static void toDot(Automaton a) throws IOException {
+ final String s = a.toDot();
+ Writer w = new OutputStreamWriter(new FileOutputStream("/x/tmp/out.dot"));
+ w.write(s);
+ w.close();
+ System.out.println("TEST: saved to /x/tmp/out.dot");
+ }
+ */
+
+ private static readonly Automaton SEP_A = BasicAutomata.MakeChar(TokenStreamToAutomaton.POS_SEP);
+ private static readonly Automaton HOLE_A = BasicAutomata.MakeChar(TokenStreamToAutomaton.HOLE);
+
+ private Automaton Join(params string[] strings)
+ {
+ IList<Automaton> @as = new List<Automaton>();
+ foreach (string s in strings)
+ {
+ @as.Add(BasicAutomata.MakeString(s));
+ @as.Add(SEP_A);
+ }
+ @as.RemoveAt(@as.Count - 1);
+ return BasicOperations.Concatenate(@as);
+ }
+
+ private Automaton Join(params Automaton[] @as)
+ {
+ return BasicOperations.Concatenate(Arrays.AsList(@as));
+ }
+
+ private Automaton S2a(string s)
+ {
+ return BasicAutomata.MakeString(s);
+ }
+
+ [Test]
+ public virtual void TestTwoTokens()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1), Token("def", 1, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton expected = Join("abc", "def");
+
+ //toDot(actual);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestHole()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1), Token("def", 2, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+
+ Automaton expected = Join(S2a("abc"), SEP_A, HOLE_A, SEP_A, S2a("def"));
+
+ //toDot(actual);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestOverlappedTokensSausage()
+ {
+ // Two tokens on top of each other (sausage):
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1), Token("xyz", 0, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton a1 = BasicAutomata.MakeString("abc");
+ Automaton a2 = BasicAutomata.MakeString("xyz");
+ Automaton expected = BasicOperations.Union(a1, a2);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestOverlappedTokensLattice()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1), Token("xyz", 0, 2), Token("def", 1, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton a1 = BasicAutomata.MakeString("xyz");
+ Automaton a2 = Join("abc", "def");
+
+ Automaton expected = BasicOperations.Union(a1, a2);
+ //toDot(actual);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestSynOverHole()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("a", 1, 1), Token("X", 0, 2), Token("b", 2, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton a1 = BasicOperations.Union(Join(S2a("a"), SEP_A, HOLE_A), BasicAutomata.MakeString("X"));
+ Automaton expected = BasicOperations.Concatenate(a1, Join(SEP_A, S2a("b")));
+ //toDot(actual);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestSynOverHole2()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("xyz", 1, 1), Token("abc", 0, 3), Token("def", 2, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton expected = BasicOperations.Union(Join(S2a("xyz"), SEP_A, HOLE_A, SEP_A, S2a("def")), BasicAutomata.MakeString("abc"));
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestOverlappedTokensLattice2()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1), Token("xyz", 0, 3), Token("def", 1, 1), Token("ghi", 1, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton a1 = BasicAutomata.MakeString("xyz");
+ Automaton a2 = Join("abc", "def", "ghi");
+ Automaton expected = BasicOperations.Union(a1, a2);
+ //toDot(actual);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ [Test]
+ public virtual void TestToDot()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 1, 1, 0, 4) });
+ StringWriter w = new StringWriter();
+ (new TokenStreamToDot("abcd", ts, (TextWriter)(w))).ToDot();
+ Assert.IsTrue(w.ToString().IndexOf("abc / abcd") != -1);
+ }
+
+ [Test]
+ public virtual void TestStartsWithHole()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("abc", 2, 1) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton expected = Join(HOLE_A, SEP_A, S2a("abc"));
+ //toDot(actual);
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+
+ // TODO: testEndsWithHole... but we need posInc to set in TS.end()
+
+ [Test]
+ public virtual void TestSynHangingOverEnd()
+ {
+ TokenStream ts = new CannedTokenStream(new Token[] { Token("a", 1, 1), Token("X", 0, 10) });
+ Automaton actual = (new TokenStreamToAutomaton()).ToAutomaton(ts);
+ Automaton expected = BasicOperations.Union(BasicAutomata.MakeString("a"), BasicAutomata.MakeString("X"));
+ Assert.IsTrue(BasicOperations.SameLanguage(expected, actual));
+ }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestLookaheadTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestLookaheadTokenFilter.cs b/src/Lucene.Net.Tests/Analysis/TestLookaheadTokenFilter.cs
new file mode 100644
index 0000000..ee733bd
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestLookaheadTokenFilter.cs
@@ -0,0 +1,129 @@
+using Lucene.Net.Attributes;
+using Lucene.Net.Randomized.Generators;
+using NUnit.Framework;
+using System;
+using System.IO;
+
+namespace Lucene.Net.Analysis
+{
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ [TestFixture]
+ public class TestLookaheadTokenFilter : BaseTokenStreamTestCase
+ {
+#if !NETSTANDARD
+ // LUCENENET: There is no Timeout on NUnit for .NET Core.
+ [Timeout(int.MaxValue)]
+#endif
+ [Test, LongRunningTest, HasTimeout]
+ public virtual void TestRandomStrings()
+ {
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper(this);
+ CheckRandomData(Random(), a, 200 * RANDOM_MULTIPLIER, 8192);
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper : Analyzer
+ {
+ private readonly TestLookaheadTokenFilter OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper(TestLookaheadTokenFilter outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Random random = Random();
+ Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, random.NextBoolean());
+ TokenStream output = new MockRandomLookaheadTokenFilter(random, tokenizer);
+ return new TokenStreamComponents(tokenizer, output);
+ }
+ }
+
+ private class NeverPeeksLookaheadTokenFilter : LookaheadTokenFilter<LookaheadTokenFilter.Position>
+ {
+ public NeverPeeksLookaheadTokenFilter(TokenStream input)
+ : base(input)
+ {
+ }
+
+ protected internal override LookaheadTokenFilter.Position NewPosition()
+ {
+ return new LookaheadTokenFilter.Position();
+ }
+
+ public sealed override bool IncrementToken()
+ {
+ return NextToken();
+ }
+ }
+
+#if !NETSTANDARD
+ // LUCENENET: There is no Timeout on NUnit for .NET Core.
+ [Timeout(int.MaxValue)]
+#endif
+ [Test, LongRunningTest, HasTimeout]
+ public virtual void TestNeverCallingPeek()
+ {
+ Analyzer a = new NCPAnalyzerAnonymousInnerClassHelper(this);
+ CheckRandomData(Random(), a, 200 * RANDOM_MULTIPLIER, 8192);
+ }
+
+ private class NCPAnalyzerAnonymousInnerClassHelper : Analyzer
+ {
+ private readonly TestLookaheadTokenFilter OuterInstance;
+
+ public NCPAnalyzerAnonymousInnerClassHelper(TestLookaheadTokenFilter outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, Random().NextBoolean());
+ TokenStream output = new NeverPeeksLookaheadTokenFilter(tokenizer);
+ return new TokenStreamComponents(tokenizer, output);
+ }
+ }
+
+ [Test]
+ public virtual void TestMissedFirstToken()
+ {
+ Analyzer analyzer = new AnalyzerAnonymousInnerClassHelper2(this);
+
+ AssertAnalyzesTo(analyzer, "Only he who is running knows .", new string[] { "Only", "Only-huh?", "he", "he-huh?", "who", "who-huh?", "is", "is-huh?", "running", "running-huh?", "knows", "knows-huh?", ".", ".-huh?" });
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper2 : Analyzer
+ {
+ private readonly TestLookaheadTokenFilter OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper2(TestLookaheadTokenFilter outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ TrivialLookaheadFilter filter = new TrivialLookaheadFilter(source);
+ return new TokenStreamComponents(source, filter);
+ }
+ }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestMockAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestMockAnalyzer.cs b/src/Lucene.Net.Tests/Analysis/TestMockAnalyzer.cs
new file mode 100644
index 0000000..cd8f315
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestMockAnalyzer.cs
@@ -0,0 +1,420 @@
+using System;
+using Lucene.Net.Attributes;
+using Lucene.Net.Documents;
+using Lucene.Net.Index;
+
+namespace Lucene.Net.Analysis
+{
+ using Lucene.Net.Randomized.Generators;
+ using Lucene.Net.Support;
+ using NUnit.Framework;
+ using System.IO;
+ using AtomicReader = Lucene.Net.Index.AtomicReader;
+ using Automaton = Lucene.Net.Util.Automaton.Automaton;
+ using AutomatonTestUtil = Lucene.Net.Util.Automaton.AutomatonTestUtil;
+ using BasicAutomata = Lucene.Net.Util.Automaton.BasicAutomata;
+ using BasicOperations = Lucene.Net.Util.Automaton.BasicOperations;
+ using BytesRef = Lucene.Net.Util.BytesRef;
+ using CharacterRunAutomaton = Lucene.Net.Util.Automaton.CharacterRunAutomaton;
+ using DocsAndPositionsEnum = Lucene.Net.Index.DocsAndPositionsEnum;
+
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ using Document = Documents.Document;
+ using Field = Field;
+ using Fields = Lucene.Net.Index.Fields;
+ using FieldType = FieldType;
+ using IOUtils = Lucene.Net.Util.IOUtils;
+ using RandomIndexWriter = Lucene.Net.Index.RandomIndexWriter;
+ using RegExp = Lucene.Net.Util.Automaton.RegExp;
+ using Terms = Lucene.Net.Index.Terms;
+ using TermsEnum = Lucene.Net.Index.TermsEnum;
+ using TestUtil = Lucene.Net.Util.TestUtil;
+
+ [TestFixture]
+ public class TestMockAnalyzer : BaseTokenStreamTestCase
+ {
+ /// <summary>
+ /// Test a configuration that behaves a lot like WhitespaceAnalyzer </summary>
+ [Test]
+ public virtual void TestWhitespace()
+ {
+ Analyzer a = new MockAnalyzer(Random());
+ AssertAnalyzesTo(a, "A bc defg hiJklmn opqrstuv wxy z ", new string[] { "a", "bc", "defg", "hijklmn", "opqrstuv", "wxy", "z" });
+ AssertAnalyzesTo(a, "aba cadaba shazam", new string[] { "aba", "cadaba", "shazam" });
+ AssertAnalyzesTo(a, "break on whitespace", new string[] { "break", "on", "whitespace" });
+ }
+
+ /// <summary>
+ /// Test a configuration that behaves a lot like SimpleAnalyzer </summary>
+ [Test]
+ public virtual void TestSimple()
+ {
+ Analyzer a = new MockAnalyzer(Random(), MockTokenizer.SIMPLE, true);
+ AssertAnalyzesTo(a, "a-bc123 defg+hijklmn567opqrstuv78wxy_z ", new string[] { "a", "bc", "defg", "hijklmn", "opqrstuv", "wxy", "z" });
+ AssertAnalyzesTo(a, "aba4cadaba-Shazam", new string[] { "aba", "cadaba", "shazam" });
+ AssertAnalyzesTo(a, "break+on/Letters", new string[] { "break", "on", "letters" });
+ }
+
+ /// <summary>
+ /// Test a configuration that behaves a lot like KeywordAnalyzer </summary>
+ [Test]
+ public virtual void TestKeyword()
+ {
+ Analyzer a = new MockAnalyzer(Random(), MockTokenizer.KEYWORD, false);
+ AssertAnalyzesTo(a, "a-bc123 defg+hijklmn567opqrstuv78wxy_z ", new string[] { "a-bc123 defg+hijklmn567opqrstuv78wxy_z " });
+ AssertAnalyzesTo(a, "aba4cadaba-Shazam", new string[] { "aba4cadaba-Shazam" });
+ AssertAnalyzesTo(a, "break+on/Nothing", new string[] { "break+on/Nothing" });
+ // currently though emits no tokens for empty string: maybe we can do it,
+ // but we don't want to emit tokens infinitely...
+ AssertAnalyzesTo(a, "", new string[0]);
+ }
+
+ // Test some regular expressions as tokenization patterns
+ /// <summary>
+ /// Test a configuration where each character is a term </summary>
+ [Test]
+ public virtual void TestSingleChar()
+ {
+ var single = new CharacterRunAutomaton((new RegExp(".")).ToAutomaton());
+ Analyzer a = new MockAnalyzer(Random(), single, false);
+ AssertAnalyzesTo(a, "foobar", new[] { "f", "o", "o", "b", "a", "r" }, new[] { 0, 1, 2, 3, 4, 5 }, new[] { 1, 2, 3, 4, 5, 6 });
+ CheckRandomData(Random(), a, 100);
+ }
+
+ /// <summary>
+ /// Test a configuration where two characters makes a term </summary>
+ [Test]
+ public virtual void TestTwoChars()
+ {
+ CharacterRunAutomaton single = new CharacterRunAutomaton((new RegExp("..")).ToAutomaton());
+ Analyzer a = new MockAnalyzer(Random(), single, false);
+ AssertAnalyzesTo(a, "foobar", new string[] { "fo", "ob", "ar" }, new int[] { 0, 2, 4 }, new int[] { 2, 4, 6 });
+ // make sure when last term is a "partial" match that End() is correct
+ AssertTokenStreamContents(a.TokenStream("bogus", new StringReader("fooba")), new string[] { "fo", "ob" }, new int[] { 0, 2 }, new int[] { 2, 4 }, new int[] { 1, 1 }, new int?(5));
+ CheckRandomData(Random(), a, 100);
+ }
+
+ /// <summary>
+ /// Test a configuration where three characters makes a term </summary>
+ [Test]
+ public virtual void TestThreeChars()
+ {
+ CharacterRunAutomaton single = new CharacterRunAutomaton((new RegExp("...")).ToAutomaton());
+ Analyzer a = new MockAnalyzer(Random(), single, false);
+ AssertAnalyzesTo(a, "foobar", new string[] { "foo", "bar" }, new int[] { 0, 3 }, new int[] { 3, 6 });
+ // make sure when last term is a "partial" match that End() is correct
+ AssertTokenStreamContents(a.TokenStream("bogus", new StringReader("fooba")), new string[] { "foo" }, new int[] { 0 }, new int[] { 3 }, new int[] { 1 }, new int?(5));
+ CheckRandomData(Random(), a, 100);
+ }
+
+ /// <summary>
+ /// Test a configuration where word starts with one uppercase </summary>
+ [Test]
+ public virtual void TestUppercase()
+ {
+ CharacterRunAutomaton single = new CharacterRunAutomaton((new RegExp("[A-Z][a-z]*")).ToAutomaton());
+ Analyzer a = new MockAnalyzer(Random(), single, false);
+ AssertAnalyzesTo(a, "FooBarBAZ", new string[] { "Foo", "Bar", "B", "A", "Z" }, new int[] { 0, 3, 6, 7, 8 }, new int[] { 3, 6, 7, 8, 9 });
+ AssertAnalyzesTo(a, "aFooBar", new string[] { "Foo", "Bar" }, new int[] { 1, 4 }, new int[] { 4, 7 });
+ CheckRandomData(Random(), a, 100);
+ }
+
+ /// <summary>
+ /// Test a configuration that behaves a lot like StopAnalyzer </summary>
+ [Test]
+ public virtual void TestStop()
+ {
+ Analyzer a = new MockAnalyzer(Random(), MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET);
+ AssertAnalyzesTo(a, "the quick brown a fox", new string[] { "quick", "brown", "fox" }, new int[] { 2, 1, 2 });
+ }
+
+ /// <summary>
+ /// Test a configuration that behaves a lot like KeepWordFilter </summary>
+ [Test]
+ public virtual void TestKeep()
+ {
+ CharacterRunAutomaton keepWords = new CharacterRunAutomaton(BasicOperations.Complement(Automaton.Union(Arrays.AsList(BasicAutomata.MakeString("foo"), BasicAutomata.MakeString("bar")))));
+ Analyzer a = new MockAnalyzer(Random(), MockTokenizer.SIMPLE, true, keepWords);
+ AssertAnalyzesTo(a, "quick foo brown bar bar fox foo", new string[] { "foo", "bar", "bar", "foo" }, new int[] { 2, 2, 1, 2 });
+ }
+
+ /// <summary>
+ /// Test a configuration that behaves a lot like LengthFilter </summary>
+ [Test]
+ public virtual void TestLength()
+ {
+ CharacterRunAutomaton length5 = new CharacterRunAutomaton((new RegExp(".{5,}")).ToAutomaton());
+ Analyzer a = new MockAnalyzer(Random(), MockTokenizer.WHITESPACE, true, length5);
+ AssertAnalyzesTo(a, "ok toolong fine notfine", new string[] { "ok", "fine" }, new int[] { 1, 2 });
+ }
+
+ /// <summary>
+ /// Test MockTokenizer encountering a too long token </summary>
+ [Test]
+ public virtual void TestTooLongToken()
+ {
+ Analyzer whitespace = new AnalyzerAnonymousInnerClassHelper(this);
+ AssertTokenStreamContents(whitespace.TokenStream("bogus", new StringReader("test 123 toolong ok ")), new string[] { "test", "123", "toolo", "ng", "ok" }, new int[] { 0, 5, 9, 14, 17 }, new int[] { 4, 8, 14, 16, 19 }, new int?(20));
+ AssertTokenStreamContents(whitespace.TokenStream("bogus", new StringReader("test 123 toolo")), new string[] { "test", "123", "toolo" }, new int[] { 0, 5, 9 }, new int[] { 4, 8, 14 }, new int?(14));
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper : Analyzer
+ {
+ private readonly TestMockAnalyzer OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper(TestMockAnalyzer outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false, 5);
+ return new TokenStreamComponents(t, t);
+ }
+ }
+
+ [Test]
+ public virtual void TestLUCENE_3042()
+ {
+ string testString = "t";
+
+ Analyzer analyzer = new MockAnalyzer(Random());
+ Exception priorException = null;
+ TokenStream stream = analyzer.TokenStream("dummy", new StringReader(testString));
+ try
+ {
+ stream.Reset();
+ while (stream.IncrementToken())
+ {
+ // consume
+ }
+ stream.End();
+ }
+ catch (Exception e)
+ {
+ priorException = e;
+ }
+ finally
+ {
+ IOUtils.CloseWhileHandlingException(priorException, stream);
+ }
+
+ AssertAnalyzesTo(analyzer, testString, new string[] { "t" });
+ }
+
+ /// <summary>
+ /// blast some random strings through the analyzer </summary>
+ [Test]
+ public virtual void TestRandomStrings()
+ {
+ CheckRandomData(Random(), new MockAnalyzer(Random()), AtLeast(1000));
+ }
+
+ /// <summary>
+ /// blast some random strings through differently configured tokenizers </summary>
+#if !NETSTANDARD
+ // LUCENENET: There is no Timeout on NUnit for .NET Core.
+ [Timeout(int.MaxValue)]
+#endif
+ [Test, LongRunningTest, HasTimeout]
+ public virtual void TestRandomRegexps()
+ {
+ int iters = AtLeast(30);
+ for (int i = 0; i < iters; i++)
+ {
+ CharacterRunAutomaton dfa = new CharacterRunAutomaton(AutomatonTestUtil.RandomAutomaton(Random()));
+ bool lowercase = Random().NextBoolean();
+ int limit = TestUtil.NextInt(Random(), 0, 500);
+ Analyzer a = new AnalyzerAnonymousInnerClassHelper2(this, dfa, lowercase, limit);
+ CheckRandomData(Random(), a, 100);
+ a.Dispose();
+ }
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper2 : Analyzer
+ {
+ private readonly TestMockAnalyzer OuterInstance;
+
+ private CharacterRunAutomaton Dfa;
+ private bool Lowercase;
+ private int Limit;
+
+ public AnalyzerAnonymousInnerClassHelper2(TestMockAnalyzer outerInstance, CharacterRunAutomaton dfa, bool lowercase, int limit)
+ {
+ this.OuterInstance = outerInstance;
+ this.Dfa = dfa;
+ this.Lowercase = lowercase;
+ this.Limit = limit;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer t = new MockTokenizer(reader, Dfa, Lowercase, Limit);
+ return new TokenStreamComponents(t, t);
+ }
+ }
+
+ [Test]
+ public virtual void TestForwardOffsets()
+ {
+ int num = AtLeast(10000);
+ for (int i = 0; i < num; i++)
+ {
+ string s = TestUtil.RandomHtmlishString(Random(), 20);
+ StringReader reader = new StringReader(s);
+ MockCharFilter charfilter = new MockCharFilter(reader, 2);
+ MockAnalyzer analyzer = new MockAnalyzer(Random());
+ Exception priorException = null;
+ TokenStream ts = analyzer.TokenStream("bogus", charfilter.m_input);
+ try
+ {
+ ts.Reset();
+ while (ts.IncrementToken())
+ {
+ ;
+ }
+ ts.End();
+ }
+ catch (Exception e)
+ {
+ priorException = e;
+ }
+ finally
+ {
+ IOUtils.CloseWhileHandlingException(priorException, ts);
+ }
+ }
+ }
+
+ [Test]
+ public virtual void TestWrapReader()
+ {
+ // LUCENE-5153: test that wrapping an analyzer's reader is allowed
+ Random random = Random();
+
+ Analyzer @delegate = new MockAnalyzer(random);
+ Analyzer a = new AnalyzerWrapperAnonymousInnerClassHelper(this, @delegate.Strategy, @delegate);
+
+ CheckOneTerm(a, "abc", "aabc");
+ }
+
+ private class AnalyzerWrapperAnonymousInnerClassHelper : AnalyzerWrapper
+ {
+ private readonly TestMockAnalyzer OuterInstance;
+
+ private Analyzer @delegate;
+
+ public AnalyzerWrapperAnonymousInnerClassHelper(TestMockAnalyzer outerInstance, ReuseStrategy getReuseStrategy, Analyzer @delegate)
+ : base(getReuseStrategy)
+ {
+ this.OuterInstance = outerInstance;
+ this.@delegate = @delegate;
+ }
+
+ protected override TextReader WrapReader(string fieldName, TextReader reader)
+ {
+ return new MockCharFilter(reader, 7);
+ }
+
+ protected override TokenStreamComponents WrapComponents(string fieldName, TokenStreamComponents components)
+ {
+ return components;
+ }
+
+ protected override Analyzer GetWrappedAnalyzer(string fieldName)
+ {
+ return @delegate;
+ }
+ }
+
+ [Test]
+ public virtual void TestChangeGaps()
+ {
+ // LUCENE-5324: check that it is possible to change the wrapper's gaps
+ int positionGap = Random().Next(1000);
+ int offsetGap = Random().Next(1000);
+ Analyzer @delegate = new MockAnalyzer(Random());
+ Analyzer a = new AnalyzerWrapperAnonymousInnerClassHelper2(this, @delegate.Strategy, positionGap, offsetGap, @delegate);
+
+ RandomIndexWriter writer = new RandomIndexWriter(Random(), NewDirectory(), Similarity, TimeZone);
+ Document doc = new Document();
+ FieldType ft = new FieldType();
+ ft.IsIndexed = true;
+ ft.IndexOptions = IndexOptions.DOCS_ONLY;
+ ft.IsTokenized = true;
+ ft.StoreTermVectors = true;
+ ft.StoreTermVectorPositions = true;
+ ft.StoreTermVectorOffsets = true;
+ doc.Add(new Field("f", "a", ft));
+ doc.Add(new Field("f", "a", ft));
+ writer.AddDocument(doc, a);
+ AtomicReader reader = GetOnlySegmentReader(writer.Reader);
+ Fields fields = reader.GetTermVectors(0);
+ Terms terms = fields.GetTerms("f");
+ TermsEnum te = terms.GetIterator(null);
+ Assert.AreEqual(new BytesRef("a"), te.Next());
+ DocsAndPositionsEnum dpe = te.DocsAndPositions(null, null);
+ Assert.AreEqual(0, dpe.NextDoc());
+ Assert.AreEqual(2, dpe.Freq);
+ Assert.AreEqual(0, dpe.NextPosition());
+ Assert.AreEqual(0, dpe.StartOffset);
+ int endOffset = dpe.EndOffset;
+ Assert.AreEqual(1 + positionGap, dpe.NextPosition());
+ Assert.AreEqual(1 + endOffset + offsetGap, dpe.EndOffset);
+ Assert.AreEqual(null, te.Next());
+ reader.Dispose();
+ writer.Dispose();
+ writer.w.Directory.Dispose();
+ }
+
+ private class AnalyzerWrapperAnonymousInnerClassHelper2 : AnalyzerWrapper
+ {
+ private readonly TestMockAnalyzer OuterInstance;
+
+ private int PositionGap;
+ private int OffsetGap;
+ private Analyzer @delegate;
+
+ public AnalyzerWrapperAnonymousInnerClassHelper2(TestMockAnalyzer outerInstance, ReuseStrategy getReuseStrategy, int positionGap, int offsetGap, Analyzer @delegate)
+ : base(getReuseStrategy)
+ {
+ this.OuterInstance = outerInstance;
+ this.PositionGap = positionGap;
+ this.OffsetGap = offsetGap;
+ this.@delegate = @delegate;
+ }
+
+ protected override Analyzer GetWrappedAnalyzer(string fieldName)
+ {
+ return @delegate;
+ }
+
+ public override int GetPositionIncrementGap(string fieldName)
+ {
+ return PositionGap;
+ }
+
+ public override int GetOffsetGap(string fieldName)
+ {
+ return OffsetGap;
+ }
+ }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestMockCharFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestMockCharFilter.cs b/src/Lucene.Net.Tests/Analysis/TestMockCharFilter.cs
new file mode 100644
index 0000000..75fc8cd
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestMockCharFilter.cs
@@ -0,0 +1,59 @@
+using NUnit.Framework;
+using System.IO;
+
+namespace Lucene.Net.Analysis
+{
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ [TestFixture]
+ public class TestMockCharFilter : BaseTokenStreamTestCase
+ {
+ [Test]
+ public virtual void Test()
+ {
+ Analyzer analyzer = new AnalyzerAnonymousInnerClassHelper(this);
+
+ AssertAnalyzesTo(analyzer, "aab", new string[] { "aab" }, new int[] { 0 }, new int[] { 3 });
+
+ AssertAnalyzesTo(analyzer, "aabaa", new string[] { "aabaa" }, new int[] { 0 }, new int[] { 5 });
+
+ AssertAnalyzesTo(analyzer, "aabcdefgaa", new string[] { "aabcdefgaa" }, new int[] { 0 }, new int[] { 10 });
+ }
+
+ private class AnalyzerAnonymousInnerClassHelper : Analyzer
+ {
+ private readonly TestMockCharFilter OuterInstance;
+
+ public AnalyzerAnonymousInnerClassHelper(TestMockCharFilter outerInstance)
+ {
+ this.OuterInstance = outerInstance;
+ }
+
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ {
+ Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
+ return new TokenStreamComponents(tokenizer, tokenizer);
+ }
+
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
+ {
+ return new MockCharFilter(reader, 7);
+ }
+ }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestNumericTokenStream.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestNumericTokenStream.cs b/src/Lucene.Net.Tests/Analysis/TestNumericTokenStream.cs
new file mode 100644
index 0000000..8c020e4
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestNumericTokenStream.cs
@@ -0,0 +1,144 @@
+using Lucene.Net.Analysis.TokenAttributes;
+
+namespace Lucene.Net.Analysis
+{
+ using NUnit.Framework;
+ using System;
+
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ using BytesRef = Lucene.Net.Util.BytesRef;
+ using CharTermAttribute = Lucene.Net.Analysis.TokenAttributes.CharTermAttribute;
+ using ICharTermAttribute = Lucene.Net.Analysis.TokenAttributes.ICharTermAttribute;
+ using NumericUtils = Lucene.Net.Util.NumericUtils;
+
+ [TestFixture]
+ public class TestNumericTokenStream : BaseTokenStreamTestCase
+ {
+ internal const long Lvalue = 4573245871874382L;
+ internal const int Ivalue = 123456;
+
+ [NUnit.Framework.Test]
+ public virtual void TestLongStream()
+ {
+ using (NumericTokenStream stream = (new NumericTokenStream()).SetInt64Value(Lvalue)) {
+ // use getAttribute to test if attributes really exist, if not an IAE will be throwed
+ ITermToBytesRefAttribute bytesAtt = stream.GetAttribute<ITermToBytesRefAttribute>();
+ ITypeAttribute typeAtt = stream.GetAttribute<ITypeAttribute>();
+ NumericTokenStream.INumericTermAttribute numericAtt = stream.GetAttribute<NumericTokenStream.INumericTermAttribute>();
+ BytesRef bytes = bytesAtt.BytesRef;
+ stream.Reset();
+ Assert.AreEqual(64, numericAtt.ValueSize);
+ for (int shift = 0; shift < 64; shift += NumericUtils.PRECISION_STEP_DEFAULT)
+ {
+ Assert.IsTrue(stream.IncrementToken(), "New token is available");
+ Assert.AreEqual(shift, numericAtt.Shift, "Shift value wrong");
+ bytesAtt.FillBytesRef();
+ Assert.AreEqual(Lvalue & ~((1L << shift) - 1L), NumericUtils.PrefixCodedToInt64(bytes), "Term is incorrectly encoded");
+ Assert.AreEqual(Lvalue & ~((1L << shift) - 1L), numericAtt.RawValue, "Term raw value is incorrectly encoded");
+ Assert.AreEqual((shift == 0) ? NumericTokenStream.TOKEN_TYPE_FULL_PREC : NumericTokenStream.TOKEN_TYPE_LOWER_PREC, typeAtt.Type, "Type incorrect");
+ }
+ Assert.IsFalse(stream.IncrementToken(), "More tokens available");
+ stream.End();
+ }
+ }
+
+ [NUnit.Framework.Test]
+ public virtual void TestIntStream()
+ {
+ NumericTokenStream stream = (new NumericTokenStream()).SetInt32Value(Ivalue);
+ // use getAttribute to test if attributes really exist, if not an IAE will be throwed
+ ITermToBytesRefAttribute bytesAtt = stream.GetAttribute<ITermToBytesRefAttribute>();
+ ITypeAttribute typeAtt = stream.GetAttribute<ITypeAttribute>();
+ NumericTokenStream.INumericTermAttribute numericAtt = stream.GetAttribute<NumericTokenStream.INumericTermAttribute>();
+ BytesRef bytes = bytesAtt.BytesRef;
+ stream.Reset();
+ Assert.AreEqual(32, numericAtt.ValueSize);
+ for (int shift = 0; shift < 32; shift += NumericUtils.PRECISION_STEP_DEFAULT)
+ {
+ Assert.IsTrue(stream.IncrementToken(), "New token is available");
+ Assert.AreEqual(shift, numericAtt.Shift, "Shift value wrong");
+ bytesAtt.FillBytesRef();
+ Assert.AreEqual(Ivalue & ~((1 << shift) - 1), NumericUtils.PrefixCodedToInt32(bytes), "Term is incorrectly encoded");
+ Assert.AreEqual(((long)Ivalue) & ~((1L << shift) - 1L), numericAtt.RawValue, "Term raw value is incorrectly encoded");
+ Assert.AreEqual((shift == 0) ? NumericTokenStream.TOKEN_TYPE_FULL_PREC : NumericTokenStream.TOKEN_TYPE_LOWER_PREC, typeAtt.Type, "Type incorrect");
+ }
+ Assert.IsFalse(stream.IncrementToken(), "More tokens available");
+ stream.End();
+ stream.Dispose();
+ }
+
+ [NUnit.Framework.Test]
+ public virtual void TestNotInitialized()
+ {
+ NumericTokenStream stream = new NumericTokenStream();
+
+ try
+ {
+ stream.Reset();
+ Assert.Fail("reset() should not succeed.");
+ }
+ catch (Exception)
+ {
+ // pass
+ }
+
+ try
+ {
+ stream.IncrementToken();
+ Assert.Fail("IncrementToken() should not succeed.");
+ }
+ catch (Exception)
+ {
+ // pass
+ }
+ }
+
+ public interface ITestAttribute : ICharTermAttribute
+ {
+ }
+
+ public class TestAttribute : CharTermAttribute, ITestAttribute
+ {
+ }
+
+ [NUnit.Framework.Test]
+ public virtual void TestCTA()
+ {
+ NumericTokenStream stream = new NumericTokenStream();
+ try
+ {
+ stream.AddAttribute<ICharTermAttribute>();
+ Assert.Fail("Succeeded to add CharTermAttribute.");
+ }
+ catch (System.ArgumentException iae)
+ {
+ Assert.IsTrue(iae.Message.StartsWith("NumericTokenStream does not support"));
+ }
+ try
+ {
+ stream.AddAttribute<ITestAttribute>();
+ Assert.Fail("Succeeded to add TestAttribute.");
+ }
+ catch (System.ArgumentException iae)
+ {
+ Assert.IsTrue(iae.Message.StartsWith("NumericTokenStream does not support"));
+ }
+ }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestPosition.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestPosition.cs b/src/Lucene.Net.Tests/Analysis/TestPosition.cs
new file mode 100644
index 0000000..54eb089
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestPosition.cs
@@ -0,0 +1,27 @@
+namespace Lucene.Net.Analysis
+{
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ /// <summary>
+ /// Trivial position class.
+ /// </summary>
+ public class TestPosition : LookaheadTokenFilter.Position
+ {
+ public virtual string Fact { get; set; }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestReusableStringReader.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestReusableStringReader.cs b/src/Lucene.Net.Tests/Analysis/TestReusableStringReader.cs
new file mode 100644
index 0000000..150a3f8
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestReusableStringReader.cs
@@ -0,0 +1,71 @@
+using System.Text;
+
+namespace Lucene.Net.Analysis
+{
+ using NUnit.Framework;
+ using LuceneTestCase = Lucene.Net.Util.LuceneTestCase;
+
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ [TestFixture]
+ public class TestReusableStringReader : LuceneTestCase
+ {
+ [Test]
+ public virtual void Test()
+ {
+ char[] buf = new char[4];
+
+ using (ReusableStringReader reader = new ReusableStringReader())
+ {
+ Assert.AreEqual(-1, reader.Read());
+ Assert.AreEqual(-1, reader.Read(new char[1], 0, 1));
+ Assert.AreEqual(-1, reader.Read(new char[2], 1, 1));
+ //Assert.AreEqual(-1, reader.Read(CharBuffer.wrap(new char[2])));
+
+ reader.SetValue("foobar");
+ Assert.AreEqual(4, reader.Read(buf, 0, 4));
+ Assert.AreEqual("foob", new string(buf));
+ Assert.AreEqual(2, reader.Read(buf, 0, 2));
+ Assert.AreEqual("ar", new string(buf, 0, 2));
+ Assert.AreEqual(-1, reader.Read(buf, 2, 0));
+ }
+
+ using (ReusableStringReader reader = new ReusableStringReader())
+ {
+ reader.SetValue("foobar");
+ Assert.AreEqual(0, reader.Read(buf, 1, 0));
+ Assert.AreEqual(3, reader.Read(buf, 1, 3));
+ Assert.AreEqual("foo", new string(buf, 1, 3));
+ Assert.AreEqual(2, reader.Read(buf, 2, 2));
+ Assert.AreEqual("ba", new string(buf, 2, 2));
+ Assert.AreEqual('r', (char)reader.Read());
+ Assert.AreEqual(-1, reader.Read(buf, 2, 0));
+ reader.Dispose();
+
+ reader.SetValue("foobar");
+ StringBuilder sb = new StringBuilder();
+ int ch;
+ while ((ch = reader.Read()) != -1)
+ {
+ sb.Append((char)ch);
+ }
+ Assert.AreEqual("foobar", sb.ToString());
+ }
+ }
+ }
+}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/96822396/src/Lucene.Net.Tests/Analysis/TestToken.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/Analysis/TestToken.cs b/src/Lucene.Net.Tests/Analysis/TestToken.cs
new file mode 100644
index 0000000..5e3fa66
--- /dev/null
+++ b/src/Lucene.Net.Tests/Analysis/TestToken.cs
@@ -0,0 +1,305 @@
+using System.Collections.Generic;
+using System.Text;
+
+namespace Lucene.Net.Analysis
+{
+ /*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+ using Lucene.Net.Analysis.TokenAttributes;
+ using NUnit.Framework;
+ using Attribute = Lucene.Net.Util.Attribute;
+ using BytesRef = Lucene.Net.Util.BytesRef;
+ using IAttribute = Lucene.Net.Util.IAttribute;
+ using LuceneTestCase = Lucene.Net.Util.LuceneTestCase;
+ using TestUtil = Lucene.Net.Util.TestUtil;
+
+ [TestFixture]
+ public class TestToken : LuceneTestCase
+ {
+ [Test]
+ public virtual void TestCtor()
+ {
+ Token t = new Token();
+ char[] content = "hello".ToCharArray();
+ t.CopyBuffer(content, 0, content.Length);
+ Assert.AreNotSame(t.Buffer, content);
+ Assert.AreEqual(0, t.StartOffset);
+ Assert.AreEqual(0, t.EndOffset);
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual("word", t.Type);
+ Assert.AreEqual(0, t.Flags);
+
+ t = new Token(6, 22);
+ t.CopyBuffer(content, 0, content.Length);
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual(6, t.StartOffset);
+ Assert.AreEqual(22, t.EndOffset);
+ Assert.AreEqual("word", t.Type);
+ Assert.AreEqual(0, t.Flags);
+
+ t = new Token(6, 22, 7);
+ t.CopyBuffer(content, 0, content.Length);
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual(6, t.StartOffset);
+ Assert.AreEqual(22, t.EndOffset);
+ Assert.AreEqual("word", t.Type);
+ Assert.AreEqual(7, t.Flags);
+
+ t = new Token(6, 22, "junk");
+ t.CopyBuffer(content, 0, content.Length);
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual("hello", t.ToString());
+ Assert.AreEqual(6, t.StartOffset);
+ Assert.AreEqual(22, t.EndOffset);
+ Assert.AreEqual("junk", t.Type);
+ Assert.AreEqual(0, t.Flags);
+ }
+
+ [Test]
+ public virtual void TestResize()
+ {
+ Token t = new Token();
+ char[] content = "hello".ToCharArray();
+ t.CopyBuffer(content, 0, content.Length);
+ for (int i = 0; i < 2000; i++)
+ {
+ t.ResizeBuffer(i);
+ Assert.IsTrue(i <= t.Buffer.Length);
+ Assert.AreEqual("hello", t.ToString());
+ }
+ }
+
+ [Test]
+ public virtual void TestGrow()
+ {
+ Token t = new Token();
+ StringBuilder buf = new StringBuilder("ab");
+ for (int i = 0; i < 20; i++)
+ {
+ char[] content = buf.ToString().ToCharArray();
+ t.CopyBuffer(content, 0, content.Length);
+ Assert.AreEqual(buf.Length, t.Length);
+ Assert.AreEqual(buf.ToString(), t.ToString());
+ buf.Append(buf.ToString());
+ }
+ Assert.AreEqual(1048576, t.Length);
+
+ // now as a string, second variant
+ t = new Token();
+ buf = new StringBuilder("ab");
+ for (int i = 0; i < 20; i++)
+ {
+ t.SetEmpty().Append(buf);
+ string content = buf.ToString();
+ Assert.AreEqual(content.Length, t.Length);
+ Assert.AreEqual(content, t.ToString());
+ buf.Append(content);
+ }
+ Assert.AreEqual(1048576, t.Length);
+
+ // Test for slow growth to a long term
+ t = new Token();
+ buf = new StringBuilder("a");
+ for (int i = 0; i < 20000; i++)
+ {
+ t.SetEmpty().Append(buf);
+ string content = buf.ToString();
+ Assert.AreEqual(content.Length, t.Length);
+ Assert.AreEqual(content, t.ToString());
+ buf.Append("a");
+ }
+ Assert.AreEqual(20000, t.Length);
+
+ // Test for slow growth to a long term
+ t = new Token();
+ buf = new StringBuilder("a");
+ for (int i = 0; i < 20000; i++)
+ {
+ t.SetEmpty().Append(buf);
+ string content = buf.ToString();
+ Assert.AreEqual(content.Length, t.Length);
+ Assert.AreEqual(content, t.ToString());
+ buf.Append("a");
+ }
+ Assert.AreEqual(20000, t.Length);
+ }
+
+ [Test]
+ public virtual void TestToString()
+ {
+ char[] b = new char[] { 'a', 'l', 'o', 'h', 'a' };
+ Token t = new Token("", 0, 5);
+ t.CopyBuffer(b, 0, 5);
+ Assert.AreEqual("aloha", t.ToString());
+
+ t.SetEmpty().Append("hi there");
+ Assert.AreEqual("hi there", t.ToString());
+ }
+
+ [Test]
+ public virtual void TestTermBufferEquals()
+ {
+ Token t1a = new Token();
+ char[] content1a = "hello".ToCharArray();
+ t1a.CopyBuffer(content1a, 0, 5);
+ Token t1b = new Token();
+ char[] content1b = "hello".ToCharArray();
+ t1b.CopyBuffer(content1b, 0, 5);
+ Token t2 = new Token();
+ char[] content2 = "hello2".ToCharArray();
+ t2.CopyBuffer(content2, 0, 6);
+ Assert.IsTrue(t1a.Equals(t1b));
+ Assert.IsFalse(t1a.Equals(t2));
+ Assert.IsFalse(t2.Equals(t1b));
+ }
+
+ [Test]
+ public virtual void TestMixedStringArray()
+ {
+ Token t = new Token("hello", 0, 5);
+ Assert.AreEqual(t.Length, 5);
+ Assert.AreEqual(t.ToString(), "hello");
+ t.SetEmpty().Append("hello2");
+ Assert.AreEqual(t.Length, 6);
+ Assert.AreEqual(t.ToString(), "hello2");
+ t.CopyBuffer("hello3".ToCharArray(), 0, 6);
+ Assert.AreEqual(t.ToString(), "hello3");
+
+ char[] buffer = t.Buffer;
+ buffer[1] = 'o';
+ Assert.AreEqual(t.ToString(), "hollo3");
+ }
+
+ [Test]
+ public virtual void TestClone()
+ {
+ Token t = new Token(0, 5);
+ char[] content = "hello".ToCharArray();
+ t.CopyBuffer(content, 0, 5);
+ char[] buf = t.Buffer;
+ Token copy = AssertCloneIsEqual(t);
+ Assert.AreEqual(t.ToString(), copy.ToString());
+ Assert.AreNotSame(buf, copy.Buffer);
+
+ BytesRef pl = new BytesRef(new byte[] { 1, 2, 3, 4 });
+ t.Payload = pl;
+ copy = AssertCloneIsEqual(t);
+ Assert.AreEqual(pl, copy.Payload);
+ Assert.AreNotSame(pl, copy.Payload);
+ }
+
+ [Test]
+ public virtual void TestCopyTo()
+ {
+ Token t = new Token();
+ Token copy = AssertCopyIsEqual(t);
+ Assert.AreEqual("", t.ToString());
+ Assert.AreEqual("", copy.ToString());
+
+ t = new Token(0, 5);
+ char[] content = "hello".ToCharArray();
+ t.CopyBuffer(content, 0, 5);
+ char[] buf = t.Buffer;
+ copy = AssertCopyIsEqual(t);
+ Assert.AreEqual(t.ToString(), copy.ToString());
+ Assert.AreNotSame(buf, copy.Buffer);
+
+ BytesRef pl = new BytesRef(new byte[] { 1, 2, 3, 4 });
+ t.Payload = pl;
+ copy = AssertCopyIsEqual(t);
+ Assert.AreEqual(pl, copy.Payload);
+ Assert.AreNotSame(pl, copy.Payload);
+ }
+
+ public interface ISenselessAttribute : Lucene.Net.Util.IAttribute
+ {
+ }
+
+ public sealed class SenselessAttribute : Attribute, ISenselessAttribute
+ {
+ public override void CopyTo(IAttribute target)
+ {
+ }
+
+ public override void Clear()
+ {
+ }
+
+ public override bool Equals(object o)
+ {
+ return (o is SenselessAttribute);
+ }
+
+ public override int GetHashCode()
+ {
+ return 0;
+ }
+ }
+
+ [Test]
+ public virtual void TestTokenAttributeFactory()
+ {
+ TokenStream ts = new MockTokenizer(Token.TOKEN_ATTRIBUTE_FACTORY, new System.IO.StringReader("foo bar"), MockTokenizer.WHITESPACE, false, MockTokenizer.DEFAULT_MAX_TOKEN_LENGTH);
+
+ Assert.IsTrue(ts.AddAttribute<ISenselessAttribute>() is SenselessAttribute, "SenselessAttribute is not implemented by SenselessAttributeImpl");
+
+ Assert.IsTrue(ts.AddAttribute<ICharTermAttribute>() is Token, "CharTermAttribute is not implemented by Token");
+ Assert.IsTrue(ts.AddAttribute<IOffsetAttribute>() is Token, "OffsetAttribute is not implemented by Token");
+ Assert.IsTrue(ts.AddAttribute<IFlagsAttribute>() is Token, "FlagsAttribute is not implemented by Token");
+ Assert.IsTrue(ts.AddAttribute<IPayloadAttribute>() is Token, "PayloadAttribute is not implemented by Token");
+ Assert.IsTrue(ts.AddAttribute<IPositionIncrementAttribute>() is Token, "PositionIncrementAttribute is not implemented by Token");
+ Assert.IsTrue(ts.AddAttribute<ITypeAttribute>() is Token, "TypeAttribute is not implemented by Token");
+ }
+
+ [Test]
+ public virtual void TestAttributeReflection()
+ {
+ Token t = new Token("foobar", 6, 22, 8);
+ TestUtil.AssertAttributeReflection(t, new Dictionary<string, object>()
+ {
+ { typeof(ICharTermAttribute).Name + "#term", "foobar" },
+ { typeof(ITermToBytesRefAttribute).Name + "#bytes", new BytesRef("foobar") },
+ { typeof(IOffsetAttribute).Name + "#startOffset", 6 },
+ { typeof(IOffsetAttribute).Name + "#endOffset", 22 },
+ { typeof(IPositionIncrementAttribute).Name + "#positionIncrement", 1 },
+ { typeof(IPayloadAttribute).Name + "#payload", null },
+ { typeof(ITypeAttribute).Name + "#type", TypeAttribute_Fields.DEFAULT_TYPE },
+ { typeof(IFlagsAttribute).Name + "#flags", 8 }
+ });
+ }
+
+ public static T AssertCloneIsEqual<T>(T att) where T : Attribute
+ {
+ T clone = (T)att.Clone();
+ Assert.AreEqual(att, clone, "Clone must be equal");
+ Assert.AreEqual(att.GetHashCode(), clone.GetHashCode(), "Clone's hashcode must be equal");
+ return clone;
+ }
+
+ public static T AssertCopyIsEqual<T>(T att) where T : Attribute
+ {
+ T copy = (T)System.Activator.CreateInstance(att.GetType());
+ att.CopyTo(copy);
+ Assert.AreEqual(att, copy, "Copied instance must be equal");
+ Assert.AreEqual(att.GetHashCode(), copy.GetHashCode(), "Copied instance's hashcode must be equal");
+ return copy;
+ }
+ }
+}
\ No newline at end of file