You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucenenet.apache.org by ni...@apache.org on 2020/02/09 06:16:03 UTC
[lucenenet] 11/35: BREAKING: Lucene.Net.Support: Marked Arrays and
Collections classes internal
This is an automated email from the ASF dual-hosted git repository.
nightowl888 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/lucenenet.git
commit bb5d6bdfe12066233ebab4ae2633e5badb517d98
Author: Shad Storhaug <sh...@shadstorhaug.com>
AuthorDate: Thu Feb 6 23:29:53 2020 +0700
BREAKING: Lucene.Net.Support: Marked Arrays and Collections classes internal
---
.../Analysis/Ar/TestArabicNormalizationFilter.cs | 2 +-
.../Analysis/Ar/TestArabicStemFilter.cs | 2 +-
.../Analysis/Bg/TestBulgarianStemmer.cs | 2 +-
.../Analysis/Br/TestBrazilianStemmer.cs | 2 +-
.../CharFilters/HTMLStripCharFilterTest.cs | 4 +--
.../Analysis/CharFilters/TestMappingCharFilter.cs | 12 ++++----
.../Analysis/Cjk/TestCJKAnalyzer.cs | 8 +++---
.../Analysis/Cjk/TestCJKBigramFilter.cs | 10 +++----
.../Analysis/Cjk/TestCJKWidthFilter.cs | 4 +--
.../Analysis/Ckb/TestSoraniNormalizationFilter.cs | 2 +-
.../Analysis/Ckb/TestSoraniStemFilter.cs | 2 +-
.../Analysis/Cn/TestChineseTokenizer.cs | 4 +--
.../Analysis/Commongrams/CommonGramsFilterTest.cs | 8 +++---
.../Compound/TestCompoundWordTokenFilter.cs | 12 ++++----
.../Analysis/Core/TestAnalyzers.cs | 4 +--
.../Analysis/Core/TestBugInSomething.cs | 8 +++---
.../Analysis/Core/TestDuelingAnalyzers.cs | 12 ++++----
.../Analysis/Core/TestFactories.cs | 4 +--
.../Analysis/Core/TestRandomChains.cs | 4 +--
.../Analysis/Core/TestStandardAnalyzer.cs | 8 +++---
.../Analysis/Core/TestStopFilter.cs | 2 +-
.../Analysis/Core/TestUAX29URLEmailTokenizer.cs | 14 +++++-----
.../Analysis/Cz/TestCzechStemmer.cs | 2 +-
.../Analysis/De/TestGermanLightStemFilter.cs | 6 ++--
.../Analysis/De/TestGermanMinimalStemFilter.cs | 6 ++--
.../Analysis/De/TestGermanNormalizationFilter.cs | 4 +--
.../Analysis/De/TestGermanStemFilter.cs | 6 ++--
.../Analysis/El/TestGreekStemmer.cs | 2 +-
.../Analysis/En/TestEnglishMinimalStemFilter.cs | 4 +--
.../Analysis/En/TestKStemmer.cs | 4 +--
.../Analysis/En/TestPorterStemFilter.cs | 4 +--
.../Analysis/Es/TestSpanishLightStemFilter.cs | 4 +--
.../Analysis/Fa/TestPersianCharFilter.cs | 4 +--
.../Analysis/Fa/TestPersianNormalizationFilter.cs | 2 +-
.../Analysis/Fi/TestFinnishLightStemFilter.cs | 6 ++--
.../Analysis/Fr/TestFrenchLightStemFilter.cs | 6 ++--
.../Analysis/Fr/TestFrenchMinimalStemFilter.cs | 6 ++--
.../Analysis/Ga/TestIrishLowerCaseFilter.cs | 2 +-
.../Analysis/Gl/TestGalicianMinimalStemFilter.cs | 6 ++--
.../Analysis/Gl/TestGalicianStemFilter.cs | 4 +--
.../Analysis/Hi/TestHindiNormalizer.cs | 2 +-
.../Analysis/Hi/TestHindiStemmer.cs | 2 +-
.../Analysis/Hu/TestHungarianLightStemFilter.cs | 6 ++--
.../Analysis/Hunspell/TestHunspellStemFilter.cs | 6 ++--
.../Analysis/Id/TestIndonesianStemmer.cs | 6 ++--
.../Analysis/In/TestIndicNormalizer.cs | 2 +-
.../Analysis/It/TestItalianLightStemFilter.cs | 4 +--
.../Analysis/Lv/TestLatvianStemmer.cs | 4 +--
.../Miscellaneous/TestASCIIFoldingFilter.cs | 4 +--
.../Miscellaneous/TestCapitalizationFilter.cs | 4 +--
.../Miscellaneous/TestCodepointCountFilter.cs | 2 +-
.../Miscellaneous/TestHyphenatedWordsFilter.cs | 4 +--
.../Analysis/Miscellaneous/TestKeepWordFilter.cs | 2 +-
.../Analysis/Miscellaneous/TestLengthFilter.cs | 2 +-
.../Miscellaneous/TestLimitTokenPositionFilter.cs | 2 +-
.../TestLucene47WordDelimiterFilter.cs | 10 +++----
.../Miscellaneous/TestPerFieldAnalyzerWrapper.cs | 4 +--
.../TestRemoveDuplicatesTokenFilter.cs | 4 +--
.../Miscellaneous/TestScandinavianFoldingFilter.cs | 4 +--
.../TestScandinavianNormalizationFilter.cs | 4 +--
.../Analysis/Miscellaneous/TestTrimFilter.cs | 6 ++--
.../Miscellaneous/TestWordDelimiterFilter.cs | 16 +++++------
.../Analysis/NGram/EdgeNGramTokenFilterTest.cs | 10 +++----
.../Analysis/NGram/EdgeNGramTokenizerTest.cs | 4 +--
.../Analysis/NGram/NGramTokenFilterTest.cs | 6 ++--
.../Analysis/NGram/NGramTokenizerTest.cs | 2 +-
.../Analysis/No/TestNorwegianLightStemFilter.cs | 8 +++---
.../Analysis/No/TestNorwegianMinimalStemFilter.cs | 8 +++---
.../Analysis/Path/TestPathHierarchyTokenizer.cs | 4 +--
.../Path/TestReversePathHierarchyTokenizer.cs | 4 +--
.../Pattern/TestPatternCaptureGroupTokenFilter.cs | 2 +-
.../Pattern/TestPatternReplaceCharFilter.cs | 4 +--
.../Analysis/Pattern/TestPatternReplaceFilter.cs | 6 ++--
.../Analysis/Pattern/TestPatternTokenizer.cs | 4 +--
.../Analysis/Pt/TestPortugueseLightStemFilter.cs | 6 ++--
.../Analysis/Pt/TestPortugueseMinimalStemFilter.cs | 6 ++--
.../Analysis/Pt/TestPortugueseStemFilter.cs | 6 ++--
.../Analysis/Reverse/TestReverseStringFilter.cs | 4 +--
.../Analysis/Ru/TestRussianLightStemFilter.cs | 6 ++--
.../Analysis/Shingle/ShingleAnalyzerWrapperTest.cs | 2 +-
.../Analysis/Shingle/ShingleFilterTest.cs | 6 ++--
.../Analysis/Snowball/TestSnowball.cs | 4 +--
.../Analysis/Snowball/TestSnowballVocab.cs | 2 +-
.../Analysis/Sv/TestSwedishLightStemFilter.cs | 6 ++--
.../Analysis/Synonym/TestSolrSynonymParser.cs | 4 +--
.../Analysis/Synonym/TestSynonymMapFilter.cs | 32 +++++++++++-----------
.../Analysis/Synonym/TestWordnetSynonymParser.cs | 2 +-
.../Analysis/Tr/TestTurkishLowerCaseFilter.cs | 2 +-
.../Analysis/Util/TestCharTokenizers.cs | 6 ++--
.../Analysis/Util/TestElision.cs | 2 +-
.../Analysis/Wikipedia/WikipediaTokenizerTest.cs | 4 +--
.../Morfologik/TestMorfologikAnalyzer.cs | 3 +-
.../Morfologik/TestMorfologikFilterFactory.cs | 2 +-
.../Analyzing/AnalyzingInfixSuggesterTest.cs | 4 +--
.../Suggest/Analyzing/AnalyzingSuggesterTest.cs | 30 ++++++++++----------
.../Suggest/Analyzing/FuzzySuggesterTest.cs | 14 +++++-----
.../Suggest/Analyzing/TestFreeTextSuggester.cs | 2 +-
src/Lucene.Net/Properties/AssemblyInfo.cs | 7 +++++
src/Lucene.Net/Support/Arrays.cs | 32 +++++++++++-----------
src/Lucene.Net/Support/Collections.cs | 2 +-
100 files changed, 287 insertions(+), 279 deletions(-)
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicNormalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicNormalizationFilter.cs
index a21aafd..3bda1ee 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicNormalizationFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicNormalizationFilter.cs
@@ -136,7 +136,7 @@ namespace Lucene.Net.Analysis.Ar
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ArabicNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicStemFilter.cs
index 4223fb2..024430f 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ar/TestArabicStemFilter.cs
@@ -194,7 +194,7 @@ namespace Lucene.Net.Analysis.Ar
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ArabicStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Bg/TestBulgarianStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Bg/TestBulgarianStemmer.cs
index 9374f71..8f4a4d6 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Bg/TestBulgarianStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Bg/TestBulgarianStemmer.cs
@@ -247,7 +247,7 @@ namespace Lucene.Net.Analysis.Bg
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new BulgarianStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Br/TestBrazilianStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Br/TestBrazilianStemmer.cs
index 213aef3..d61a6b1 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Br/TestBrazilianStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Br/TestBrazilianStemmer.cs
@@ -187,7 +187,7 @@ namespace Lucene.Net.Analysis.Br
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new BrazilianStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/HTMLStripCharFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/HTMLStripCharFilterTest.cs
index 04877ee..7f5eef2 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/HTMLStripCharFilterTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/HTMLStripCharFilterTest.cs
@@ -41,13 +41,13 @@ namespace Lucene.Net.Analysis.CharFilters
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, tokenizer);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new HTMLStripCharFilter(reader);
}
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/TestMappingCharFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/TestMappingCharFilter.cs
index 131eb3c..7c2e877 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/TestMappingCharFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/CharFilters/TestMappingCharFilter.cs
@@ -228,13 +228,13 @@ namespace Lucene.Net.Analysis.CharFilters
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, tokenizer);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new MappingCharFilter(outerInstance.normMap, reader);
}
@@ -269,13 +269,13 @@ namespace Lucene.Net.Analysis.CharFilters
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, tokenizer);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new MappingCharFilter(map, reader);
}
@@ -307,13 +307,13 @@ namespace Lucene.Net.Analysis.CharFilters
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, tokenizer);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new MappingCharFilter(map, reader);
}
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKAnalyzer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKAnalyzer.cs
index 0ed7077..e47344c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKAnalyzer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKAnalyzer.cs
@@ -157,13 +157,13 @@ namespace Lucene.Net.Analysis.Cjk
this.norm = norm;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, new CJKBigramFilter(tokenizer));
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new MappingCharFilter(norm, reader);
}
@@ -211,7 +211,7 @@ namespace Lucene.Net.Analysis.Cjk
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenFilter filter = new FakeStandardTokenizer(tokenizer);
@@ -254,7 +254,7 @@ namespace Lucene.Net.Analysis.Cjk
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new CJKBigramFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKBigramFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKBigramFilter.cs
index da5662d..e40ee6e 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKBigramFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKBigramFilter.cs
@@ -32,7 +32,7 @@ namespace Lucene.Net.Analysis.Cjk
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(t, new CJKBigramFilter(t));
@@ -47,7 +47,7 @@ namespace Lucene.Net.Analysis.Cjk
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(t, new CJKBigramFilter(t, (CJKScript)0xff, true));
@@ -76,7 +76,7 @@ namespace Lucene.Net.Analysis.Cjk
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(t, new CJKBigramFilter(t, CJKScript.HAN));
@@ -99,7 +99,7 @@ namespace Lucene.Net.Analysis.Cjk
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(t, new CJKBigramFilter(t, (CJKScript)0xff, false));
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Cjk
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(t, new CJKBigramFilter(t, CJKScript.HAN, true));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKWidthFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKWidthFilter.cs
index 0164c58..9ef17fc 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKWidthFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cjk/TestCJKWidthFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.Cjk
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new CJKWidthFilter(source));
@@ -85,7 +85,7 @@ namespace Lucene.Net.Analysis.Cjk
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new CJKWidthFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniNormalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniNormalizationFilter.cs
index 7256e7d..56fec10 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniNormalizationFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniNormalizationFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.Ckb
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new SoraniNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniStemFilter.cs
index 3457947..b053b88 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ckb/TestSoraniStemFilter.cs
@@ -110,7 +110,7 @@ namespace Lucene.Net.Analysis.Ckb
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new SoraniStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cn/TestChineseTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cn/TestChineseTokenizer.cs
index dca54c6..c70de77 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cn/TestChineseTokenizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cn/TestChineseTokenizer.cs
@@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.Cn
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
return new TokenStreamComponents(new ChineseTokenizer(reader));
}
@@ -89,7 +89,7 @@ namespace Lucene.Net.Analysis.Cn
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new WhitespaceTokenizer(LuceneVersion.LUCENE_CURRENT, reader);
return new TokenStreamComponents(tokenizer, new ChineseFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Commongrams/CommonGramsFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Commongrams/CommonGramsFilterTest.cs
index f69545f..f2c965c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Commongrams/CommonGramsFilterTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Commongrams/CommonGramsFilterTest.cs
@@ -139,7 +139,7 @@ namespace Lucene.Net.Analysis.CommonGrams
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new CommonGramsQueryFilter(new CommonGramsFilter(TEST_VERSION_CURRENT, tokenizer, commonWords)));
@@ -197,7 +197,7 @@ namespace Lucene.Net.Analysis.CommonGrams
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new CommonGramsFilter(TEST_VERSION_CURRENT, tokenizer, commonWords));
@@ -305,7 +305,7 @@ namespace Lucene.Net.Analysis.CommonGrams
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
CommonGramsFilter cgf = new CommonGramsFilter(TEST_VERSION_CURRENT, t, commonWords);
@@ -323,7 +323,7 @@ namespace Lucene.Net.Analysis.CommonGrams
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
CommonGramsFilter cgf = new CommonGramsFilter(TEST_VERSION_CURRENT, t, commonWords);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Compound/TestCompoundWordTokenFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Compound/TestCompoundWordTokenFilter.cs
index 30ee3a6..b13be6d 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Compound/TestCompoundWordTokenFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Compound/TestCompoundWordTokenFilter.cs
@@ -262,14 +262,14 @@ namespace Lucene.Net.Analysis.Compound
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenFilter filter = new DictionaryCompoundWordTokenFilter(TEST_VERSION_CURRENT, tokenizer, dict);
return new TokenStreamComponents(tokenizer, filter);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new MappingCharFilter(normMap, reader);
}
@@ -306,7 +306,7 @@ namespace Lucene.Net.Analysis.Compound
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new DictionaryCompoundWordTokenFilter(TEST_VERSION_CURRENT, tokenizer, dict));
@@ -326,7 +326,7 @@ namespace Lucene.Net.Analysis.Compound
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenFilter filter = new HyphenationCompoundWordTokenFilter(TEST_VERSION_CURRENT, tokenizer, hyphenator);
@@ -364,7 +364,7 @@ namespace Lucene.Net.Analysis.Compound
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new DictionaryCompoundWordTokenFilter(TEST_VERSION_CURRENT, tokenizer, dict));
@@ -384,7 +384,7 @@ namespace Lucene.Net.Analysis.Compound
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
TokenFilter filter = new HyphenationCompoundWordTokenFilter(TEST_VERSION_CURRENT, tokenizer, hyphenator);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestAnalyzers.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestAnalyzers.cs
index a8ee1b1..b26f4b4 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestAnalyzers.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestAnalyzers.cs
@@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Core
private class LowerCaseWhitespaceAnalyzer : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new WhitespaceTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, new LowerCaseFilter(TEST_VERSION_CURRENT, tokenizer));
@@ -126,7 +126,7 @@ namespace Lucene.Net.Analysis.Core
private class UpperCaseWhitespaceAnalyzer : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new WhitespaceTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, new UpperCaseFilter(TEST_VERSION_CURRENT, tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestBugInSomething.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestBugInSomething.cs
index 53f8267..cb26701 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestBugInSomething.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestBugInSomething.cs
@@ -66,14 +66,14 @@ namespace Lucene.Net.Analysis.Core
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new MockTokenizer(new TestRandomChains.CheckThatYouDidntReadAnythingReaderWrapper(reader), MockTokenFilter.ENGLISH_STOPSET, false, -65);
TokenFilter f = new CommonGramsFilter(TEST_VERSION_CURRENT, t, cas);
return new TokenStreamComponents(t, f);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
reader = new MockCharFilter(reader, 0);
reader = new MappingCharFilter(map, reader);
@@ -322,7 +322,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new EdgeNGramTokenizer(TEST_VERSION_CURRENT, reader, 2, 94);
//TokenStream stream = new SopTokenFilter(tokenizer);
@@ -357,7 +357,7 @@ namespace Lucene.Net.Analysis.Core
this.table = table;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new WikipediaTokenizer(reader);
TokenStream stream = new SopTokenFilter(tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestDuelingAnalyzers.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestDuelingAnalyzers.cs
index 350a826..34581a6 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestDuelingAnalyzers.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestDuelingAnalyzers.cs
@@ -78,7 +78,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -111,7 +111,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -140,7 +140,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -172,7 +172,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -201,7 +201,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -233,7 +233,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestFactories.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestFactories.cs
index 28a11dd..4b428c5 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestFactories.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestFactories.cs
@@ -212,7 +212,7 @@ namespace Lucene.Net.Analysis.Core
this.tokenfilter = tokenfilter;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tf = tokenizer.Create(reader);
if (tokenfilter != null)
@@ -226,7 +226,7 @@ namespace Lucene.Net.Analysis.Core
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
if (charFilter != null)
{
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestRandomChains.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestRandomChains.cs
index 14550a6..586c11d 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestRandomChains.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestRandomChains.cs
@@ -851,7 +851,7 @@ namespace Lucene.Net.Analysis.Core
}
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Random random = new Random(seed);
TokenizerSpec tokenizerSpec = NewTokenizer(random, reader);
@@ -861,7 +861,7 @@ namespace Lucene.Net.Analysis.Core
return new TokenStreamComponents(tokenizerSpec.tokenizer, filterSpec.stream);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
Random random = new Random(seed);
CharFilterSpec charfilterspec = NewCharFilterChain(random, reader);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStandardAnalyzer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStandardAnalyzer.cs
index b4c2d54..8b6dca9 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStandardAnalyzer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStandardAnalyzer.cs
@@ -49,7 +49,7 @@ namespace Lucene.Net.Analysis.Core
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer);
@@ -336,7 +336,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new StandardTokenizer(LuceneVersion.LUCENE_36, reader);
@@ -365,7 +365,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new StandardTokenizer(LuceneVersion.LUCENE_40, reader);
@@ -408,7 +408,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
TokenStream tokenStream = new MockGraphTokenFilter(Random, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStopFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStopFilter.cs
index 8e10148..00e2549 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStopFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestStopFilter.cs
@@ -226,7 +226,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenFilter filter = new MockSynonymFilter(outerInstance, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestUAX29URLEmailTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestUAX29URLEmailTokenizer.cs
index 1b7c7c5..0e846b2 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestUAX29URLEmailTokenizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Core/TestUAX29URLEmailTokenizer.cs
@@ -52,7 +52,7 @@ namespace Lucene.Net.Analysis.Core
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new UAX29URLEmailTokenizer(TEST_VERSION_CURRENT, reader);
@@ -119,7 +119,7 @@ namespace Lucene.Net.Analysis.Core
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
UAX29URLEmailTokenizer tokenizer = new UAX29URLEmailTokenizer(TEST_VERSION_CURRENT, reader);
tokenizer.MaxTokenLength = int.MaxValue; // Tokenize arbitrary length URLs
@@ -136,7 +136,7 @@ namespace Lucene.Net.Analysis.Core
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
UAX29URLEmailTokenizer tokenizer = new UAX29URLEmailTokenizer(TEST_VERSION_CURRENT, reader);
TokenFilter filter = new EmailFilter(tokenizer);
@@ -587,7 +587,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new UAX29URLEmailTokenizer(LuceneVersion.LUCENE_31, reader);
@@ -615,7 +615,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new UAX29URLEmailTokenizer(LuceneVersion.LUCENE_34, reader);
@@ -642,7 +642,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new UAX29URLEmailTokenizer(LuceneVersion.LUCENE_36, reader);
@@ -671,7 +671,7 @@ namespace Lucene.Net.Analysis.Core
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new UAX29URLEmailTokenizer(LuceneVersion.LUCENE_40, reader);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cz/TestCzechStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cz/TestCzechStemmer.cs
index 8a83dc0..2bc754a 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cz/TestCzechStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Cz/TestCzechStemmer.cs
@@ -312,7 +312,7 @@ namespace Lucene.Net.Analysis.Cz
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new CzechStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanLightStemFilter.cs
index 1bfe698..4f70192 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.De
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new GermanLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.De
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.De
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GermanLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanMinimalStemFilter.cs
index fcb7077..ccac752 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanMinimalStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanMinimalStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.De
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new GermanMinimalStemFilter(source));
@@ -78,7 +78,7 @@ namespace Lucene.Net.Analysis.De
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -118,7 +118,7 @@ namespace Lucene.Net.Analysis.De
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GermanMinimalStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanNormalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanNormalizationFilter.cs
index 4737fda..bbb74e4 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanNormalizationFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanNormalizationFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.De
{
}
- protected override TokenStreamComponents CreateComponents(string field, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream stream = new GermanNormalizationFilter(tokenizer);
@@ -94,7 +94,7 @@ namespace Lucene.Net.Analysis.De
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GermanNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanStemFilter.cs
index 52b59d3..d9db6d4 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/De/TestGermanStemFilter.cs
@@ -39,7 +39,7 @@ namespace Lucene.Net.Analysis.De
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new KeywordTokenizer(reader);
return new TokenStreamComponents(t, new GermanStemFilter(new LowerCaseFilter(TEST_VERSION_CURRENT, t)));
@@ -82,7 +82,7 @@ namespace Lucene.Net.Analysis.De
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -114,7 +114,7 @@ namespace Lucene.Net.Analysis.De
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GermanStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/El/TestGreekStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/El/TestGreekStemmer.cs
index 92df565..8127f55 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/El/TestGreekStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/El/TestGreekStemmer.cs
@@ -553,7 +553,7 @@ namespace Lucene.Net.Analysis.El
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GreekStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestEnglishMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestEnglishMinimalStemFilter.cs
index ec10c0d..2332cb4 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestEnglishMinimalStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestEnglishMinimalStemFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.En
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new EnglishMinimalStemFilter(source));
@@ -81,7 +81,7 @@ namespace Lucene.Net.Analysis.En
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new EnglishMinimalStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestKStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestKStemmer.cs
index 0752480..d3cf691 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestKStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestKStemmer.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.En
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true);
return new TokenStreamComponents(tokenizer, new KStemFilter(tokenizer));
@@ -76,7 +76,7 @@ namespace Lucene.Net.Analysis.En
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new KStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestPorterStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestPorterStemFilter.cs
index 52f3067..b5bc842 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestPorterStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/En/TestPorterStemFilter.cs
@@ -35,7 +35,7 @@ namespace Lucene.Net.Analysis.En
public AnalyzerAnonymousInnerClassHelper()
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new MockTokenizer(reader, MockTokenizer.KEYWORD, false);
return new TokenStreamComponents(t, new PorterStemFilter(t));
@@ -79,7 +79,7 @@ namespace Lucene.Net.Analysis.En
private class AnalyzerAnonymousInnerClassHelper2 : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new PorterStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Es/TestSpanishLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Es/TestSpanishLightStemFilter.cs
index 5da6300..5e6c023 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Es/TestSpanishLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Es/TestSpanishLightStemFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.Es
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new SpanishLightStemFilter(source));
@@ -73,7 +73,7 @@ namespace Lucene.Net.Analysis.Es
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new SpanishLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianCharFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianCharFilter.cs
index d29ac6d..17a7828 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianCharFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianCharFilter.cs
@@ -30,12 +30,12 @@ namespace Lucene.Net.Analysis.Fa
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
return new TokenStreamComponents(new MockTokenizer(reader));
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new PersianCharFilter(reader);
}
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianNormalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianNormalizationFilter.cs
index 99a9b13..8431df5 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianNormalizationFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fa/TestPersianNormalizationFilter.cs
@@ -90,7 +90,7 @@ namespace Lucene.Net.Analysis.Fa
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new PersianNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fi/TestFinnishLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fi/TestFinnishLightStemFilter.cs
index ddaba58..7e8f8dc 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fi/TestFinnishLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fi/TestFinnishLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Fi
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new FinnishLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Fi
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Fi
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new FinnishLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchLightStemFilter.cs
index b437454..f474f01 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Fr
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new FrenchLightStemFilter(source));
@@ -205,7 +205,7 @@ namespace Lucene.Net.Analysis.Fr
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -237,7 +237,7 @@ namespace Lucene.Net.Analysis.Fr
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new FrenchLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchMinimalStemFilter.cs
index 03be805..8b41071 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchMinimalStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Fr/TestFrenchMinimalStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Fr
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new FrenchMinimalStemFilter(source));
@@ -80,7 +80,7 @@ namespace Lucene.Net.Analysis.Fr
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -120,7 +120,7 @@ namespace Lucene.Net.Analysis.Fr
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new FrenchMinimalStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ga/TestIrishLowerCaseFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ga/TestIrishLowerCaseFilter.cs
index c674f58..fb54eda 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ga/TestIrishLowerCaseFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ga/TestIrishLowerCaseFilter.cs
@@ -54,7 +54,7 @@ namespace Lucene.Net.Analysis.Ga
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new IrishLowerCaseFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianMinimalStemFilter.cs
index a1e3de4..e675dc8 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianMinimalStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianMinimalStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Gl
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new GalicianMinimalStemFilter(tokenizer));
@@ -79,7 +79,7 @@ namespace Lucene.Net.Analysis.Gl
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -111,7 +111,7 @@ namespace Lucene.Net.Analysis.Gl
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GalicianMinimalStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianStemFilter.cs
index a867d8d..d70094c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Gl/TestGalicianStemFilter.cs
@@ -35,7 +35,7 @@ namespace Lucene.Net.Analysis.Gl
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source);
@@ -68,7 +68,7 @@ namespace Lucene.Net.Analysis.Gl
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new GalicianStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiNormalizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiNormalizer.cs
index 5a4b59d..baf430c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiNormalizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiNormalizer.cs
@@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Hi
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new HindiNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiStemmer.cs
index 08cf43a..26b2990 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hi/TestHindiStemmer.cs
@@ -108,7 +108,7 @@ namespace Lucene.Net.Analysis.Hi
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new HindiStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
index 89ef132..eeeef95 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Hu
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new HungarianLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Hu
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -95,7 +95,7 @@ namespace Lucene.Net.Analysis.Hu
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new HungarianLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
index 6144dea..6a310d5 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
@@ -98,7 +98,7 @@ namespace Lucene.Net.Analysis.Hunspell
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, dictionary));
@@ -121,7 +121,7 @@ namespace Lucene.Net.Analysis.Hunspell
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, dictionary));
@@ -158,7 +158,7 @@ namespace Lucene.Net.Analysis.Hunspell
this.d = d;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, d));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
index e55eff0..73857df 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
@@ -35,7 +35,7 @@ namespace Lucene.Net.Analysis.Id
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new IndonesianStemFilter(tokenizer));
@@ -129,7 +129,7 @@ namespace Lucene.Net.Analysis.Id
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new IndonesianStemFilter(tokenizer, false));
@@ -171,7 +171,7 @@ namespace Lucene.Net.Analysis.Id
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new IndonesianStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
index 47731bf..8294ffe 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
@@ -65,7 +65,7 @@ namespace Lucene.Net.Analysis.In
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new IndicNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
index d054687..b7c3e0f 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.It
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new ItalianLightStemFilter(source));
@@ -73,7 +73,7 @@ namespace Lucene.Net.Analysis.It
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ItalianLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
index 5192510..52adc34 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.Lv
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new LatvianStemFilter(tokenizer));
@@ -307,7 +307,7 @@ namespace Lucene.Net.Analysis.Lv
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new LatvianStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
index e75bf23..727af62 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
@@ -243,7 +243,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new ASCIIFoldingFilter(tokenizer, Random.nextBoolean()));
@@ -266,7 +266,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ASCIIFoldingFilter(tokenizer, Random.nextBoolean()));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
index eb5bfdc..2786d9a 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
@@ -108,7 +108,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new CapitalizationFilter(tokenizer));
@@ -131,7 +131,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new CapitalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
index 56c9189..0903925 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
@@ -50,7 +50,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new CodepointCountFilter(TEST_VERSION_CURRENT, tokenizer, 0, 5));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
index 59e2d59..d2cfd34 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
@@ -77,7 +77,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new HyphenatedWordsFilter(tokenizer));
@@ -100,7 +100,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new HyphenatedWordsFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
index 98da330..9bebbe1 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
@@ -87,7 +87,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.words = words;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream stream = new KeepWordFilter(TEST_VERSION_CURRENT, tokenizer, new CharArraySet(TEST_VERSION_CURRENT, words, true));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
index 7180aed..3c55dc8 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
@@ -60,7 +60,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new LengthFilter(TEST_VERSION_CURRENT, tokenizer, 0, 5));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
index c3afa22..3b4c014 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
@@ -55,7 +55,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.consumeAll = consumeAll;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
// if we are consuming all tokens, we can use the checks, otherwise we can't
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
index 61909bf..9eb56a9 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
@@ -293,7 +293,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protWords = protWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(tokenizer, flags, protWords));
@@ -314,7 +314,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protWords = protWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(new LargePosIncTokenFilter(outerInstance, tokenizer), flags, protWords));
@@ -335,7 +335,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protWords = protWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
StopFilter filter = new StopFilter(TEST_VERSION_CURRENT, tokenizer, StandardAnalyzer.STOP_WORDS_SET);
@@ -381,7 +381,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protectedWords = protectedWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(tokenizer, flags, protectedWords));
@@ -425,7 +425,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protectedWords = protectedWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(tokenizer, flags, protectedWords));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
index 7cca581..4eb9ac1 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
@@ -110,12 +110,12 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
return new TokenStreamComponents(new MockTokenizer(reader));
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new MockCharFilter(reader, 7);
}
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
index 7044c97..f0b4f67 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
@@ -170,7 +170,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.ignoreCase = ignoreCase;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
TokenStream stream = new SynonymFilter(tokenizer, map, ignoreCase);
@@ -194,7 +194,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new RemoveDuplicatesTokenFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
index e5cf9df..a1abfb6 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
@@ -31,7 +31,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream stream = new ScandinavianFoldingFilter(tokenizer);
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ScandinavianFoldingFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
index b7eda30..8843d06 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
@@ -33,7 +33,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream stream = new ScandinavianNormalizationFilter(tokenizer);
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ScandinavianNormalizationFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
index 9af849d..bd00b2d 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
@@ -125,7 +125,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.KEYWORD, false);
return new TokenStreamComponents(tokenizer, new TrimFilter(LuceneVersion.LUCENE_43, tokenizer, true));
@@ -141,7 +141,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.KEYWORD, false);
return new TokenStreamComponents(tokenizer, new TrimFilter(TEST_VERSION_CURRENT, tokenizer, false));
@@ -164,7 +164,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
bool updateOffsets = Random.nextBoolean();
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
index 7325027..dd4942c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
@@ -288,7 +288,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protWords = protWords;
}
- protected override TokenStreamComponents CreateComponents(string field, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protWords));
@@ -309,7 +309,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protWords = protWords;
}
- protected override TokenStreamComponents CreateComponents(string field, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, new LargePosIncTokenFilter(outerInstance, tokenizer), flags, protWords));
@@ -330,7 +330,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protWords = protWords;
}
- protected override TokenStreamComponents CreateComponents(string field, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
StopFilter filter = new StopFilter(TEST_VERSION_CURRENT, tokenizer, StandardAnalyzer.STOP_WORDS_SET);
@@ -370,7 +370,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.flags = flags;
}
- protected override TokenStreamComponents CreateComponents(string field, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, null));
@@ -410,7 +410,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.flags = flags;
}
- protected override TokenStreamComponents CreateComponents(string field, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, null));
@@ -454,7 +454,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protectedWords = protectedWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords));
@@ -500,7 +500,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords));
@@ -544,7 +544,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
this.protectedWords = protectedWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
index e037cd7..2ab3a03 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
@@ -241,7 +241,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenFilter filters = new ASCIIFoldingFilter(tokenizer);
@@ -284,7 +284,7 @@ namespace Lucene.Net.Analysis.NGram
this.max = max;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, min, max));
@@ -300,7 +300,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
#pragma warning disable 612, 618
@@ -329,7 +329,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
#pragma warning disable 612, 618
@@ -347,7 +347,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
#pragma warning disable 612, 618
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
index 662982e..db81493 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
@@ -165,7 +165,7 @@ namespace Lucene.Net.Analysis.NGram
this.max = max;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, Reader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, Reader reader)
{
Tokenizer tokenizer = new EdgeNGramTokenizer(TEST_VERSION_CURRENT, reader, min, max);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -181,7 +181,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, Reader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, Reader reader)
{
#pragma warning disable 612, 618
Tokenizer tokenizer = new Lucene43EdgeNGramTokenizer(Version.LUCENE_43, reader, Lucene43EdgeNGramTokenizer.Side.BACK, 2, 4);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
index e3bd331..df00989 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
@@ -143,7 +143,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenFilter filters = new ASCIIFoldingFilter(tokenizer);
@@ -180,7 +180,7 @@ namespace Lucene.Net.Analysis.NGram
this.max = max;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new NGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, min, max));
@@ -204,7 +204,7 @@ namespace Lucene.Net.Analysis.NGram
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new NGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, 2, 15));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
index b18a583..102e18c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
@@ -134,7 +134,7 @@ namespace Lucene.Net.Analysis.NGram
this.max = max;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new NGramTokenizer(TEST_VERSION_CURRENT, reader, min, max);
return new TokenStreamComponents(tokenizer, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
index 6b2d382..8241dfe 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.No
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new NorwegianLightStemFilter(source, NorwegianStandard.BOKMAAL));
@@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.No
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new NorwegianLightStemFilter(source, NorwegianStandard.NYNORSK));
@@ -97,7 +97,7 @@ namespace Lucene.Net.Analysis.No
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -130,7 +130,7 @@ namespace Lucene.Net.Analysis.No
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new NorwegianLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
index d8a5a9e..7482f48 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.No
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new NorwegianMinimalStemFilter(source, NorwegianStandard.BOKMAAL));
@@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.No
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new NorwegianMinimalStemFilter(source, NorwegianStandard.NYNORSK));
@@ -97,7 +97,7 @@ namespace Lucene.Net.Analysis.No
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -130,7 +130,7 @@ namespace Lucene.Net.Analysis.No
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new NorwegianMinimalStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
index 0345440..d71fe93 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
@@ -168,7 +168,7 @@ namespace Lucene.Net.Analysis.Path
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new PathHierarchyTokenizer(reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -194,7 +194,7 @@ namespace Lucene.Net.Analysis.Path
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new PathHierarchyTokenizer(reader);
return new TokenStreamComponents(tokenizer, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
index 5de3e51..7bd28b6 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
@@ -137,7 +137,7 @@ namespace Lucene.Net.Analysis.Path
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new ReversePathHierarchyTokenizer(reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -163,7 +163,7 @@ namespace Lucene.Net.Analysis.Path
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new ReversePathHierarchyTokenizer(reader);
return new TokenStreamComponents(tokenizer, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
index 667daaa..23ee77b 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
@@ -192,7 +192,7 @@ namespace Lucene.Net.Analysis.Pattern
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new PatternCaptureGroupTokenFilter(tokenizer, false, new Regex("((..)(..))", RegexOptions.Compiled)));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
index d025592..402f11d 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
@@ -282,12 +282,12 @@ namespace Lucene.Net.Analysis.Pattern
this.replacement = replacement;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, tokenizer);
}
- protected override TextReader InitReader(string fieldName, TextReader reader)
+ protected internal override TextReader InitReader(string fieldName, TextReader reader)
{
return new PatternReplaceCharFilter(p, replacement, reader);
}
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
index 8ac2195..f39cc33 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
@@ -87,7 +87,7 @@ namespace Lucene.Net.Analysis.Pattern
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream filter = new PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", false);
@@ -104,7 +104,7 @@ namespace Lucene.Net.Analysis.Pattern
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream filter = new PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", true);
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Pattern
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", true));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
index a1c4ca1..48a82e0 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
@@ -136,7 +136,7 @@ namespace Lucene.Net.Analysis.Pattern
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new PatternTokenizer(reader, new Regex("a", RegexOptions.Compiled), -1);
return new TokenStreamComponents(tokenizer);
@@ -152,7 +152,7 @@ namespace Lucene.Net.Analysis.Pattern
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new PatternTokenizer(reader, new Regex("a", RegexOptions.Compiled), 0);
return new TokenStreamComponents(tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
index f135370..f4ac8e7 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source);
@@ -111,7 +111,7 @@ namespace Lucene.Net.Analysis.Pt
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -143,7 +143,7 @@ namespace Lucene.Net.Analysis.Pt
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new PortugueseLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
index 542f427..a7a8069 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source);
@@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Pt
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Pt
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new PortugueseMinimalStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
index c384343..a64070e 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader);
TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source);
@@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Pt
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Pt
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new PortugueseStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
index 40943d3..da8ed49 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
@@ -125,7 +125,7 @@ namespace Lucene.Net.Analysis.Reverse
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new ReverseStringFilter(TEST_VERSION_CURRENT, tokenizer));
@@ -148,7 +148,7 @@ namespace Lucene.Net.Analysis.Reverse
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ReverseStringFilter(TEST_VERSION_CURRENT, tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
index dc8b045..535d398 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Ru
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new RussianLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Ru
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Ru
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new RussianLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
index 9e1fee8..9803346 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
@@ -233,7 +233,7 @@ namespace Lucene.Net.Analysis.Shingle
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
CharArraySet stopSet = StopFilter.MakeStopSet(TEST_VERSION_CURRENT, "into");
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
index 1b0f419..14a2231 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
@@ -524,7 +524,7 @@ namespace Lucene.Net.Analysis.Shingle
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer));
@@ -550,7 +550,7 @@ namespace Lucene.Net.Analysis.Shingle
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer));
@@ -573,7 +573,7 @@ namespace Lucene.Net.Analysis.Shingle
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
index be3a7e0..c6c2981 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
@@ -186,7 +186,7 @@ namespace Lucene.Net.Analysis.Snowball
this.lang = lang;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new SnowballFilter(tokenizer, lang));
@@ -220,7 +220,7 @@ namespace Lucene.Net.Analysis.Snowball
this.snowballLanguage = snowballLanguage;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new MockTokenizer(reader);
return new TokenStreamComponents(t, new SnowballFilter(t, snowballLanguage));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
index c86297b..a7ddcc5 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
@@ -89,7 +89,7 @@ namespace Lucene.Net.Analysis.Snowball
this.snowballLanguage = snowballLanguage;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer t = new KeywordTokenizer(reader);
return new TokenStreamComponents(t, new SnowballFilter(t, snowballLanguage));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
index d50df88..5e92996 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Sv
{
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(source, new SwedishLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Sv
this.exclusionSet = exclusionSet;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet);
@@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Sv
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new SwedishLightStemFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
index ea87286..9a5155e 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
@@ -63,7 +63,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -150,7 +150,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.KEYWORD, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, false));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
index 0975a51..8fe021a 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
@@ -188,7 +188,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, false));
@@ -228,7 +228,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, false));
@@ -607,7 +607,7 @@ namespace Lucene.Net.Analysis.Synonym
this.ignoreCase = ignoreCase;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, ignoreCase));
@@ -682,7 +682,7 @@ namespace Lucene.Net.Analysis.Synonym
this.ignoreCase = ignoreCase;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
TokenStream syns = new SynonymFilter(tokenizer, map, ignoreCase);
@@ -727,7 +727,7 @@ namespace Lucene.Net.Analysis.Synonym
this.ignoreCase = ignoreCase;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, ignoreCase));
@@ -777,7 +777,7 @@ namespace Lucene.Net.Analysis.Synonym
this.ignoreCase = ignoreCase;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, ignoreCase));
@@ -816,7 +816,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -898,7 +898,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -931,7 +931,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -964,7 +964,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -995,7 +995,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -1028,7 +1028,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -1097,7 +1097,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -1128,7 +1128,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -1160,7 +1160,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
@@ -1191,7 +1191,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
index fca1ce8..fa66d0a 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
@@ -57,7 +57,7 @@ namespace Lucene.Net.Analysis.Synonym
this.map = map;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, false));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
index a49a557..e2cab3f 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
@@ -86,7 +86,7 @@ namespace Lucene.Net.Analysis.Tr
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new TurkishLowerCaseFilter(tokenizer));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
index 841e388..9b6fcd6 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
@@ -152,7 +152,7 @@ namespace Lucene.Net.Analysis.Util
private sealed class AnalyzerAnonymousInnerClassHelper : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizerAnonymousInnerClassHelper(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -215,7 +215,7 @@ namespace Lucene.Net.Analysis.Util
private sealed class AnalyzerAnonymousInnerClassHelper2 : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new LetterTokenizerAnonymousInnerClassHelper2(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -261,7 +261,7 @@ namespace Lucene.Net.Analysis.Util
public AnalyzerAnonymousInnerClassHelper3()
{ }
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new NumberAndSurrogatePairTokenizer(TEST_VERSION_CURRENT, reader);
return new TokenStreamComponents(tokenizer, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
index 81c5b1d..2bc4ab3 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Util
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new KeywordTokenizer(reader);
return new TokenStreamComponents(tokenizer, new ElisionFilter(tokenizer, FrenchAnalyzer.DEFAULT_ARTICLES));
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
index b68d444..745c21c 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
@@ -127,7 +127,7 @@ namespace Lucene.Net.Analysis.Wikipedia
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new WikipediaTokenizer(reader);
return new TokenStreamComponents(tokenizer, tokenizer);
@@ -153,7 +153,7 @@ namespace Lucene.Net.Analysis.Wikipedia
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new WikipediaTokenizer(reader);
return new TokenStreamComponents(tokenizer, tokenizer);
diff --git a/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikAnalyzer.cs b/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikAnalyzer.cs
index 21cab5e..e44b1ca 100644
--- a/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikAnalyzer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikAnalyzer.cs
@@ -1,4 +1,5 @@
// Lucene version compatibility level 8.2.0
+using J2N.Text;
using Lucene.Net.Analysis.Miscellaneous;
using Lucene.Net.Analysis.Morfologik.TokenAttributes;
using Lucene.Net.Analysis.Standard;
@@ -87,7 +88,7 @@ namespace Lucene.Net.Analysis.Morfologik
ICharTermAttribute charTerm = ts.GetAttribute<ICharTermAttribute>();
while (ts.IncrementToken())
{
- Console.WriteLine(charTerm.ToString() + " => " + Collections.ToString(attribute.Tags));
+ Console.WriteLine(charTerm.ToString() + " => " + string.Format(StringFormatter.InvariantCulture, "{0}", attribute.Tags));
}
ts.End();
}
diff --git a/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikFilterFactory.cs b/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikFilterFactory.cs
index de6a886..c639872 100644
--- a/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikFilterFactory.cs
+++ b/src/Lucene.Net.Tests.Analysis.Morfologik/Morfologik/TestMorfologikFilterFactory.cs
@@ -53,7 +53,7 @@ namespace Lucene.Net.Analysis.Morfologik
public void TestDefaultDictionary()
{
StringReader reader = new StringReader("rowery bilety");
- MorfologikFilterFactory factory = new MorfologikFilterFactory(Collections.EmptyMap<String, String>());
+ MorfologikFilterFactory factory = new MorfologikFilterFactory(new Dictionary<string, string>());
factory.Inform(new ForbidResourcesLoader());
TokenStream stream = new MockTokenizer(reader); //whitespaceMockTokenizer(reader);
stream = factory.Create(stream);
diff --git a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingInfixSuggesterTest.cs b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingInfixSuggesterTest.cs
index 39cf327..bcd2b64 100644
--- a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingInfixSuggesterTest.cs
+++ b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingInfixSuggesterTest.cs
@@ -431,7 +431,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.stopWords = stopWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
MockTokenizer tokens = new MockTokenizer(reader);
return new TokenStreamComponents(tokens,
@@ -449,7 +449,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.stopWords = stopWords;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
MockTokenizer tokens = new MockTokenizer(reader);
return new TokenStreamComponents(tokens,
diff --git a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingSuggesterTest.cs b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingSuggesterTest.cs
index dcfb48b..744e365 100644
--- a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingSuggesterTest.cs
+++ b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/AnalyzingSuggesterTest.cs
@@ -313,7 +313,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -326,7 +326,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestGraphDupsTokenStreamComponents(outerInstance, tokenizer);
@@ -396,7 +396,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -410,7 +410,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestInputPathRequiredTokenStreamComponents(outerInstance, tokenizer);
@@ -506,7 +506,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -518,7 +518,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new UsualTokenStreamComponents(outerInstance, tokenizer);
@@ -737,7 +737,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.numStopChars = numStopChars;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
MockTokenizer tokenizer = new MockTokenizer(factory, reader, MockTokenizer.WHITESPACE, false, MockTokenizer.DEFAULT_MAX_TOKEN_LENGTH);
tokenizer.EnableChecks = (true);
@@ -1195,7 +1195,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -1207,7 +1207,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
{
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestDupSurfaceFormsMissingResultsTokenStreamComponents(outerInstance, tokenizer);
@@ -1283,14 +1283,14 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
internal class TestDupSurfaceFormsMissingResults2Analyzer : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestDupSurfaceFormsMissingResults2TokenStreamComponents(tokenizer);
@@ -1370,14 +1370,14 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
internal class Test0ByteKeysAnalyzer : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new Test0ByteKeysTokenStreamComponents(tokenizer);
@@ -1443,13 +1443,13 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
internal class TestTooManyExpressionsAnalyzer : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestTooManyExpressionsTokenStreamComponents(tokenizer);
diff --git a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/FuzzySuggesterTest.cs b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/FuzzySuggesterTest.cs
index f27dcf5..d1db892 100644
--- a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/FuzzySuggesterTest.cs
+++ b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/FuzzySuggesterTest.cs
@@ -266,7 +266,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -279,7 +279,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestGraphDupsTokenStreamComponents(outerInstance, tokenizer);
@@ -357,7 +357,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -371,7 +371,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new TestInputPathRequiredTokenStreamComponents(outerInstance, tokenizer);
@@ -462,7 +462,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
}
}
- protected override void SetReader(TextReader reader)
+ protected internal override void SetReader(TextReader reader)
{
}
}
@@ -474,7 +474,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.outerInstance = outerInstance;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
return new UsualTokenStreamComponents(outerInstance, tokenizer);
@@ -684,7 +684,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
this.numStopChars = numStopChars;
}
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false, MockTokenizer.DEFAULT_MAX_TOKEN_LENGTH);
tokenizer.EnableChecks = (true);
diff --git a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/TestFreeTextSuggester.cs b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/TestFreeTextSuggester.cs
index ee8db65..a956ece 100644
--- a/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/TestFreeTextSuggester.cs
+++ b/src/Lucene.Net.Tests.Suggest/Suggest/Analyzing/TestFreeTextSuggester.cs
@@ -286,7 +286,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
internal class TestEndingHoleAnalyzer : Analyzer
{
- protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+ protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
{
Tokenizer tokenizer = new MockTokenizer(reader);
CharArraySet stopSet = StopFilter.MakeStopSet(TEST_VERSION_CURRENT, "of");
diff --git a/src/Lucene.Net/Properties/AssemblyInfo.cs b/src/Lucene.Net/Properties/AssemblyInfo.cs
index cc24a9d..db154e2 100644
--- a/src/Lucene.Net/Properties/AssemblyInfo.cs
+++ b/src/Lucene.Net/Properties/AssemblyInfo.cs
@@ -67,9 +67,16 @@ using System.Runtime.CompilerServices;
[assembly: InternalsVisibleTo("Lucene.Net.TestFramework.MSTest, PublicKey=" + AssemblyKeys.PublicKey)]
[assembly: InternalsVisibleTo("Lucene.Net.TestFramework.NUnit, PublicKey=" + AssemblyKeys.PublicKey)]
[assembly: InternalsVisibleTo("Lucene.Net.TestFramework.xUnit, PublicKey=" + AssemblyKeys.PublicKey)]
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Analysis.Common, PublicKey=" + AssemblyKeys.PublicKey)]
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Expressions, PublicKey=" + AssemblyKeys.PublicKey)]
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Facet, PublicKey=" + AssemblyKeys.PublicKey)]
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Grouping, PublicKey=" + AssemblyKeys.PublicKey)]
[assembly: InternalsVisibleTo("Lucene.Net.Tests.ICU, PublicKey=" + AssemblyKeys.PublicKey)] // For Analysis.Util.TestSegmentingTokenizerBase
[assembly: InternalsVisibleTo("Lucene.Net.Tests.Misc, PublicKey=" + AssemblyKeys.PublicKey)]
[assembly: InternalsVisibleTo("Lucene.Net.Tests.QueryParser, PublicKey=" + AssemblyKeys.PublicKey)]
[assembly: InternalsVisibleTo("Lucene.Net.Tests.Cli, PublicKey=" + AssemblyKeys.PublicKey)] // For lucene-cli
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Replicator, PublicKey=" + AssemblyKeys.PublicKey)]
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Spatial, PublicKey=" + AssemblyKeys.PublicKey)]
+[assembly: InternalsVisibleTo("Lucene.Net.Tests.Suggest, PublicKey=" + AssemblyKeys.PublicKey)]
diff --git a/src/Lucene.Net/Support/Arrays.cs b/src/Lucene.Net/Support/Arrays.cs
index 7208bc5..8959472 100644
--- a/src/Lucene.Net/Support/Arrays.cs
+++ b/src/Lucene.Net/Support/Arrays.cs
@@ -6,23 +6,23 @@ using System.Text;
namespace Lucene.Net.Support
{
/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
- public static class Arrays
+ internal static class Arrays
{
/// <summary>
/// Compares the entire members of one array whith the other one.
diff --git a/src/Lucene.Net/Support/Collections.cs b/src/Lucene.Net/Support/Collections.cs
index 215504f..6370d32 100644
--- a/src/Lucene.Net/Support/Collections.cs
+++ b/src/Lucene.Net/Support/Collections.cs
@@ -28,7 +28,7 @@ namespace Lucene.Net.Support
* limitations under the License.
*/
- public static class Collections
+ internal static class Collections
{
public static IList<T> EmptyList<T>()
{