You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucenenet.apache.org by sy...@apache.org on 2016/10/02 14:36:02 UTC

[44/50] [abbrv] lucenenet git commit: Fixed most compiler warnings in Analysis (a few lingering issues).

Fixed most compiler warnings in Analysis (a few lingering issues).


Project: http://git-wip-us.apache.org/repos/asf/lucenenet/repo
Commit: http://git-wip-us.apache.org/repos/asf/lucenenet/commit/da2ae4c8
Tree: http://git-wip-us.apache.org/repos/asf/lucenenet/tree/da2ae4c8
Diff: http://git-wip-us.apache.org/repos/asf/lucenenet/diff/da2ae4c8

Branch: refs/heads/master
Commit: da2ae4c8ec48cc5f18a85b9630367daadd2cce5a
Parents: d47b608
Author: Shad Storhaug <sh...@shadstorhaug.com>
Authored: Mon Sep 19 14:55:01 2016 +0700
Committer: Shad Storhaug <sh...@shadstorhaug.com>
Committed: Sun Oct 2 17:45:07 2016 +0700

----------------------------------------------------------------------
 .../Analysis/Ar/ArabicAnalyzer.cs               |  6 +-
 .../Analysis/Br/BrazilianAnalyzer.cs            |  2 +
 .../Analysis/Ca/CatalanAnalyzer.cs              |  4 ++
 .../Analysis/CharFilter/HTMLStripCharFilter.cs  | 29 +++++---
 .../Analysis/Cjk/CJKAnalyzer.cs                 |  4 ++
 .../Analysis/Cjk/CJKBigramFilter.cs             | 10 +--
 .../Analysis/Ckb/SoraniAnalyzer.cs              |  2 +
 .../Compound/CompoundWordTokenFilterBase.cs     |  2 +
 .../Analysis/Core/StopAnalyzer.cs               |  7 +-
 .../Analysis/Core/TypeTokenFilterFactory.cs     |  2 +
 .../Analysis/Cz/CzechAnalyzer.cs                |  6 +-
 .../Analysis/Da/DanishAnalyzer.cs               |  2 +
 .../Analysis/De/GermanAnalyzer.cs               |  9 ++-
 .../Analysis/El/GreekAnalyzer.cs                |  2 +
 .../Analysis/El/GreekStemmer.cs                 | 70 ++++++++++++++++----
 .../Analysis/En/EnglishAnalyzer.cs              |  2 +
 .../Analysis/En/EnglishPossessiveFilter.cs      |  6 +-
 .../Analysis/En/KStemmer.cs                     |  2 +
 .../Analysis/Es/SpanishAnalyzer.cs              |  4 ++
 .../Analysis/Fa/PersianAnalyzer.cs              |  9 ++-
 .../Analysis/Fi/FinnishAnalyzer.cs              |  2 +
 .../Analysis/Fr/FrenchAnalyzer.cs               | 20 +++++-
 .../Analysis/Ga/IrishAnalyzer.cs                | 14 +++-
 .../Analysis/Gl/GalicianAnalyzer.cs             |  2 +
 .../Analysis/Hi/HindiAnalyzer.cs                |  2 +
 .../Analysis/Hu/HungarianAnalyzer.cs            |  4 +-
 .../Analysis/Hunspell/Stemmer.cs                |  5 +-
 .../Analysis/It/ItalianAnalyzer.cs              | 14 +++-
 .../Analysis/Lv/LatvianAnalyzer.cs              |  4 +-
 .../Miscellaneous/KeepWordFilterFactory.cs      |  2 +
 .../Miscellaneous/LengthFilterFactory.cs        |  2 +
 .../RemoveDuplicatesTokenFilter.cs              |  5 +-
 .../Analysis/Miscellaneous/TrimFilter.cs        |  2 +
 .../Analysis/Miscellaneous/TrimFilterFactory.cs |  2 +
 .../Miscellaneous/WordDelimiterFilterFactory.cs |  5 +-
 .../Analysis/Ngram/EdgeNGramFilterFactory.cs    |  2 +
 .../Analysis/Ngram/EdgeNGramTokenFilter.cs      | 14 ++--
 .../Analysis/Ngram/EdgeNGramTokenizerFactory.cs |  4 ++
 .../Ngram/Lucene43EdgeNGramTokenizer.cs         | 10 +--
 .../Analysis/Ngram/NGramTokenFilter.cs          | 10 ++-
 .../Analysis/Ngram/NGramTokenizer.cs            |  7 +-
 .../Analysis/Ngram/NGramTokenizerFactory.cs     |  4 ++
 .../Analysis/Nl/DutchAnalyzer.cs                | 23 +++++--
 .../Analysis/No/NorwegianAnalyzer.cs            |  2 +
 .../Pattern/PatternReplaceCharFilterFactory.cs  |  7 +-
 .../Analysis/Pt/PortugueseAnalyzer.cs           |  4 ++
 .../Analysis/Pt/RSLPStemmerBase.cs              |  6 +-
 .../Analysis/Reverse/ReverseStringFilter.cs     |  3 +
 .../Analysis/Ru/RussianAnalyzer.cs              | 11 ++-
 .../Analysis/Standard/ClassicTokenizerImpl.cs   |  7 +-
 .../Analysis/Standard/StandardFilter.cs         |  2 +
 .../Analysis/Standard/StandardTokenizer.cs      |  4 ++
 .../Analysis/Standard/StandardTokenizerImpl.cs  |  5 +-
 .../Standard/Std31/StandardTokenizerImpl31.cs   |  5 +-
 .../Std31/UAX29URLEmailTokenizerImpl31.cs       |  5 +-
 .../Standard/Std34/StandardTokenizerImpl34.cs   |  5 +-
 .../Std34/UAX29URLEmailTokenizerImpl34.cs       |  5 +-
 .../Std36/UAX29URLEmailTokenizerImpl36.cs       |  5 +-
 .../Standard/Std40/StandardTokenizerImpl40.cs   |  5 +-
 .../Std40/UAX29URLEmailTokenizerImpl40.cs       |  5 +-
 .../Analysis/Standard/UAX29URLEmailTokenizer.cs |  2 +
 .../Standard/UAX29URLEmailTokenizerImpl.cs      |  5 +-
 .../Analysis/Sv/SwedishAnalyzer.cs              |  2 +
 .../Analysis/Synonym/FSTSynonymFilterFactory.cs |  2 +
 .../Analysis/Synonym/SynonymFilterFactory.cs    |  4 ++
 .../Analysis/Synonym/SynonymMap.cs              |  2 +
 .../Analysis/Th/ThaiAnalyzer.cs                 | 10 ++-
 .../Analysis/Util/AbstractAnalysisFactory.cs    |  6 +-
 .../Analysis/Util/CharArrayMap.cs               |  4 +-
 .../Analysis/Util/CharacterUtils.cs             |  2 +
 .../Analysis/Util/FilteringTokenFilter.cs       |  5 +-
 .../Analysis/Util/StopwordAnalyzerBase.cs       |  6 +-
 .../Wikipedia/WikipediaTokenizerImpl.cs         |  5 +-
 .../Tartarus/Snowball/Ext/KpStemmer.cs          |  4 +-
 .../Tartarus/Snowball/Ext/PorterStemmer.cs      |  2 +-
 .../Tartarus/Snowball/SnowballProgram.cs        | 10 +--
 .../Ar/TestArabicNormalizationFilter.cs         |  2 +
 .../Analysis/Ar/TestArabicStemFilter.cs         |  4 ++
 .../Core/TestAllAnalyzersHaveFactories.cs       |  2 +
 .../Analysis/Core/TestAnalyzers.cs              |  2 +
 .../Analysis/Core/TestClassicAnalyzer.cs        |  9 ++-
 .../Analysis/Core/TestRandomChains.cs           | 16 +++++
 .../Analysis/Core/TestStandardAnalyzer.cs       |  4 ++
 .../Analysis/Core/TestStopFilter.cs             |  4 ++
 .../Analysis/Core/TestTypeTokenFilter.cs        | 12 +++-
 .../Analysis/Core/TestUAX29URLEmailTokenizer.cs |  9 ++-
 .../Analysis/De/TestGermanAnalyzer.cs           |  2 +
 .../Fa/TestPersianNormalizationFilter.cs        |  2 +
 .../Analysis/Fr/TestFrenchAnalyzer.cs           |  2 +
 .../Analysis/It/TestItalianAnalyzer.cs          |  2 +
 .../Miscellaneous/PatternAnalyzerTest.cs        |  2 +
 .../Miscellaneous/TestKeepWordFilter.cs         |  2 +
 .../Analysis/Miscellaneous/TestLengthFilter.cs  |  2 +
 .../Miscellaneous/TestLengthFilterFactory.cs    |  5 +-
 .../TestLucene47WordDelimiterFilter.cs          |  2 +-
 .../Analysis/Miscellaneous/TestTrimFilter.cs    |  3 +-
 .../Analysis/Ngram/EdgeNGramTokenFilterTest.cs  | 32 +++++++++
 .../Analysis/Ngram/EdgeNGramTokenizerTest.cs    |  8 +++
 .../Analysis/Ngram/NGramTokenFilterTest.cs      |  2 +
 .../Analysis/Ngram/TestNGramFilters.cs          |  4 ++
 .../Analysis/Nl/TestDutchStemmer.cs             |  4 ++
 .../Analysis/Position/PositionFilterTest.cs     | 21 ++++--
 .../Analysis/Snowball/TestSnowball.cs           |  3 +-
 .../Synonym/TestSynonymFilterFactory.cs         |  4 ++
 .../Analysis/Synonym/TestSynonymMapFilter.cs    |  2 +
 .../Analysis/Th/TestThaiAnalyzer.cs             |  8 ++-
 .../Analysis/Util/TestCharArraySet.cs           |  2 +
 .../Analysis/Util/TestCharacterUtils.cs         |  8 +++
 108 files changed, 578 insertions(+), 104 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicAnalyzer.cs
index 452eb1a..97f8baa 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicAnalyzer.cs
@@ -140,7 +140,11 @@ namespace Lucene.Net.Analysis.Ar
         ///         if a stem exclusion set is provided and <seealso cref="ArabicStemFilter"/>. </returns>
         public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
-            Tokenizer source = matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ? new StandardTokenizer(matchVersion, reader) : (Tokenizer)new ArabicLetterTokenizer(matchVersion, reader);
+#pragma warning disable 612, 618
+            Tokenizer source = matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) 
+                ? new StandardTokenizer(matchVersion, reader) 
+                : (Tokenizer)new ArabicLetterTokenizer(matchVersion, reader);
+#pragma warning restore 612, 618
             TokenStream result = new LowerCaseFilter(matchVersion, source);
             // the order here is important: the stopword list is not normalized!
             result = new StopFilter(matchVersion, result, stopwords);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Br/BrazilianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Br/BrazilianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Br/BrazilianAnalyzer.cs
index ed01a4a..abf84a4 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Br/BrazilianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Br/BrazilianAnalyzer.cs
@@ -65,7 +65,9 @@ namespace Lucene.Net.Analysis.Br
                     DEFAULT_STOP_SET = WordlistLoader.GetWordSet(
                         IOUtils.GetDecodingReader(typeof(BrazilianAnalyzer), typeof(BrazilianAnalyzer).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
                         "#",
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ca/CatalanAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ca/CatalanAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ca/CatalanAnalyzer.cs
index 392797b..3e0123b 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ca/CatalanAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ca/CatalanAnalyzer.cs
@@ -50,7 +50,9 @@ namespace Lucene.Net.Analysis.Ca
         public const string DEFAULT_STOPWORD_FILE = "stopwords.txt";
 
         private static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(
+#pragma warning disable 612, 618
             new CharArraySet(LuceneVersion.LUCENE_CURRENT, Arrays.AsList("d", "l", "m", "n", "s", "t"), true));
+#pragma warning restore 612, 618
 
         /// <summary>
         /// Returns an unmodifiable instance of the default stop words set. </summary>
@@ -134,7 +136,9 @@ namespace Lucene.Net.Analysis.Ca
         {
             Tokenizer source = new StandardTokenizer(matchVersion, reader);
             TokenStream result = new StandardFilter(matchVersion, source);
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
+#pragma warning restore 612, 618
             {
                 result = new ElisionFilter(result, DEFAULT_ARTICLES);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/HTMLStripCharFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/HTMLStripCharFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/HTMLStripCharFilter.cs
index cab9206..894bfcc 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/HTMLStripCharFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/HTMLStripCharFilter.cs
@@ -30650,6 +30650,7 @@ namespace Lucene.Net.Analysis.CharFilters
         /** the number of characters up to the start of the matched text */
         private int yychar;
 
+#pragma warning disable 169, 414
         /**
          * the number of characters from the last newline up to the start of the 
          * matched text
@@ -30661,6 +30662,8 @@ namespace Lucene.Net.Analysis.CharFilters
          */
         private bool zzAtBOL = true;
 
+#pragma warning restore 169, 414
+
         /** zzAtEOF == true <=> the scanner is at the EOF */
         private bool zzAtEOF;
 
@@ -30680,7 +30683,9 @@ namespace Lucene.Net.Analysis.CharFilters
             };
 
         private static readonly CharArrayMap<char> entityValues
+#pragma warning disable 612, 618
             = new CharArrayMap<char>(LuceneVersion.LUCENE_CURRENT, 253, false);
+#pragma warning restore 612, 618
 
         static HTMLStripCharFilter()
         {
@@ -30836,7 +30841,9 @@ namespace Lucene.Net.Analysis.CharFilters
                     {
                         if (null == this.escapedTags)
                         {
+#pragma warning disable 612, 618
                             this.escapedTags = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 16, true);
+#pragma warning restore 612, 618
                         }
                         this.escapedTags.Add(tag);
                     }
@@ -31138,7 +31145,7 @@ namespace Lucene.Net.Analysis.CharFilters
             {
                 message = ZZ_ERROR_MSG[errorCode];
             }
-            catch (IndexOutOfRangeException e)
+            catch (IndexOutOfRangeException /*e*/)
             {
                 message = ZZ_ERROR_MSG[ZZ_UNKNOWN_ERROR];
             }
@@ -31377,7 +31384,7 @@ namespace Lucene.Net.Analysis.CharFilters
                                 {
                                     codePoint = int.Parse(decimalCharRef, CultureInfo.InvariantCulture);
                                 }
-                                catch (Exception e)
+                                catch (Exception /*e*/)
                                 {
                                     Debug.Assert(false, "Exception parsing code point '" + decimalCharRef + "'");
                                 }
@@ -31637,7 +31644,7 @@ namespace Lucene.Net.Analysis.CharFilters
                                 {
                                     codePoint = int.Parse(hexCharRef, NumberStyles.HexNumber, CultureInfo.InvariantCulture);
                                 }
-                                catch (Exception e)
+                                catch (Exception /*e*/)
                                 {
                                     Debug.Assert(false, "Exception parsing hex code point '" + hexCharRef + "'");
                                 }
@@ -31900,7 +31907,7 @@ namespace Lucene.Net.Analysis.CharFilters
                             {
                                 highSurrogate = (char)int.Parse(surrogatePair.Substring(2, 6 - 2), NumberStyles.HexNumber, CultureInfo.InvariantCulture);
                             }
-                            catch (Exception e)
+                            catch (Exception /*e*/)
                             { // should never happen
                                 Debug.Assert(false, "Exception parsing high surrogate '"
                                             + surrogatePair.Substring(2, 6 - 2) + "'");
@@ -31909,7 +31916,7 @@ namespace Lucene.Net.Analysis.CharFilters
                             {
                                 outputSegment.UnsafeWrite((char)int.Parse(surrogatePair.Substring(10, 14 - 10), NumberStyles.HexNumber, CultureInfo.InvariantCulture));
                             }
-                            catch (Exception e)
+                            catch (Exception /*e*/)
                             { // should never happen
                                 Debug.Assert(false, "Exception parsing low surrogate '" + surrogatePair.Substring(10, 14 - 10) + "'");
                             }
@@ -31931,7 +31938,7 @@ namespace Lucene.Net.Analysis.CharFilters
                             {
                                 highSurrogate = (char)int.Parse(surrogatePair.Substring(2, 6 - 2), NumberStyles.HexNumber, CultureInfo.InvariantCulture);
                             }
-                            catch (Exception e)
+                            catch (Exception /*e*/)
                             { // should never happen
                                 Debug.Assert(false, "Exception parsing high surrogate '"
                                             + surrogatePair.Substring(2, 6 - 2) + "'");
@@ -31940,7 +31947,7 @@ namespace Lucene.Net.Analysis.CharFilters
                             { // Low surrogates are in decimal range [56320, 57343]
                                 lowSurrogate = (char)int.Parse(surrogatePair.Substring(9, 14 - 9), CultureInfo.InvariantCulture);
                             }
-                            catch (Exception e)
+                            catch (Exception /*e*/)
                             { // should never happen
                                 Debug.Assert(false, "Exception parsing low surrogate '"
                                             + surrogatePair.Substring(9, 14 - 9) + "'");
@@ -31972,7 +31979,7 @@ namespace Lucene.Net.Analysis.CharFilters
                             { // High surrogates are in decimal range [55296, 56319]
                                 highSurrogate = (char)int.Parse(surrogatePair.Substring(1, 6 - 1), CultureInfo.InvariantCulture);
                             }
-                            catch (Exception e)
+                            catch (Exception /*e*/)
                             { // should never happen
                                 Debug.Assert(false, "Exception parsing high surrogate '"
                                             + surrogatePair.Substring(1, 6 - 1) + "'");
@@ -31985,7 +31992,7 @@ namespace Lucene.Net.Analysis.CharFilters
                                 {
                                     outputSegment.UnsafeWrite((char)int.Parse(surrogatePair.Substring(10, 14 - 10), NumberStyles.HexNumber, CultureInfo.InvariantCulture));
                                 }
-                                catch (Exception e)
+                                catch (Exception /*e*/)
                                 { // should never happen
                                     Debug.Assert(false, "Exception parsing low surrogate '"
                                                 + surrogatePair.Substring(10, 14 - 10) + "'");
@@ -32012,7 +32019,7 @@ namespace Lucene.Net.Analysis.CharFilters
                             { // High surrogates are in decimal range [55296, 56319]
                                 highSurrogate = (char)int.Parse(surrogatePair.Substring(1, 6 - 1), CultureInfo.InvariantCulture);
                             }
-                            catch (Exception e)
+                            catch (Exception /*e*/)
                             { // should never happen
                                 Debug.Assert(false, "Exception parsing high surrogate '"
                                             + surrogatePair.Substring(1, 6 - 1) + "'");
@@ -32024,7 +32031,7 @@ namespace Lucene.Net.Analysis.CharFilters
                                 { // Low surrogates are in decimal range [56320, 57343]
                                     lowSurrogate = (char)int.Parse(surrogatePair.Substring(9, 14 - 9), CultureInfo.InvariantCulture);
                                 }
-                                catch (Exception e)
+                                catch (Exception /*e*/)
                                 { // should never happen
                                     Debug.Assert(false, "Exception parsing low surrogate '"
                                                 + surrogatePair.Substring(9, 14 - 9) + "'");

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKAnalyzer.cs
index 40847c3..6936e0e 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKAnalyzer.cs
@@ -93,7 +93,9 @@ namespace Lucene.Net.Analysis.Cjk
 
         public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
+#pragma warning restore 612, 618
             {
                 Tokenizer source = new StandardTokenizer(matchVersion, reader);
                 // run the widthfilter first before bigramming, it sometimes combines characters.
@@ -104,7 +106,9 @@ namespace Lucene.Net.Analysis.Cjk
             }
             else
             {
+#pragma warning disable 612, 618
                 Tokenizer source = new CJKTokenizer(reader);
+#pragma warning restore 612, 618
                 return new TokenStreamComponents(source, new StopFilter(matchVersion, source, stopwords));
             }
         }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKBigramFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKBigramFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKBigramFilter.cs
index a916aef..ebec456 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKBigramFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKBigramFilter.cs
@@ -71,13 +71,13 @@ namespace Lucene.Net.Analysis.Cjk
         private static readonly string HANGUL_TYPE = StandardTokenizer.TOKEN_TYPES[StandardTokenizer.HANGUL];
 
         // sentinel value for ignoring a script 
-        private static readonly object NO = new object();
+        private static readonly string NO = "<NO>";
 
         // these are set to either their type or NO if we want to pass them thru
-        private readonly object doHan;
-        private readonly object doHiragana;
-        private readonly object doKatakana;
-        private readonly object doHangul;
+        private readonly string doHan;
+        private readonly string doHiragana;
+        private readonly string doKatakana;
+        private readonly string doHangul;
 
         // true if we should output unigram tokens always
         private readonly bool outputUnigrams;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ckb/SoraniAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ckb/SoraniAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ckb/SoraniAnalyzer.cs
index 8abfad5..e5c0e71 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ckb/SoraniAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ckb/SoraniAnalyzer.cs
@@ -62,7 +62,9 @@ namespace Lucene.Net.Analysis.Ckb
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetWordSet(
                         IOUtils.GetDecodingReader(typeof(SoraniAnalyzer), typeof(SoraniAnalyzer).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Compound/CompoundWordTokenFilterBase.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Compound/CompoundWordTokenFilterBase.cs b/src/Lucene.Net.Analysis.Common/Analysis/Compound/CompoundWordTokenFilterBase.cs
index 4731b79..6f4b439 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Compound/CompoundWordTokenFilterBase.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Compound/CompoundWordTokenFilterBase.cs
@@ -174,7 +174,9 @@ namespace Lucene.Net.Analysis.Compound
                 int startOff = outerInstance.offsetAtt.StartOffset();
                 int endOff = outerInstance.offsetAtt.EndOffset();
 
+#pragma warning disable 612, 618
                 if (outerInstance.matchVersion.OnOrAfter(LuceneVersion.LUCENE_44) || endOff - startOff != outerInstance.termAtt.Length)
+#pragma warning restore 612, 618
                 {
                     // if length by start + end offsets doesn't match the term text then assume
                     // this is a synonym and don't adjust the offsets.

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Core/StopAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Core/StopAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Core/StopAnalyzer.cs
index f7c0cc4..a8a9af8 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Core/StopAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Core/StopAnalyzer.cs
@@ -47,8 +47,13 @@ namespace Lucene.Net.Analysis.Core
 
         static StopAnalyzer()
         {
-            IList<string> stopWords = Arrays.AsList("a", "an", "and", "are", "as", "at", "be", "but", "by", "for", "if", "in", "into", "is", "it", "no", "not", "of", "on", "or", "such", "that", "the", "their", "then", "there", "these", "they", "this", "to", "was", "will", "with");
+            IList<string> stopWords = Arrays.AsList("a", "an", "and", "are", "as", "at", "be", 
+                "but", "by", "for", "if", "in", "into", "is", "it", "no", "not", "of", "on", 
+                "or", "such", "that", "the", "their", "then", "there", "these", "they", "this", 
+                "to", "was", "will", "with");
+#pragma warning disable 612, 618
             var stopSet = new CharArraySet(LuceneVersion.LUCENE_CURRENT, stopWords, false);
+#pragma warning restore 612, 618
             ENGLISH_STOP_WORDS_SET = CharArraySet.UnmodifiableSet(stopSet);
         }
 

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Core/TypeTokenFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Core/TypeTokenFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Core/TypeTokenFilterFactory.cs
index d782ee4..78a091e 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Core/TypeTokenFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Core/TypeTokenFilterFactory.cs
@@ -86,7 +86,9 @@ namespace Lucene.Net.Analysis.Core
 
         public override TokenStream Create(TokenStream input)
         {
+#pragma warning disable 612, 618
             TokenStream filter = new TypeTokenFilter(luceneMatchVersion, enablePositionIncrements, input, stopTypes, useWhitelist);
+#pragma warning restore 612, 618
             return filter;
         }
     }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Cz/CzechAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Cz/CzechAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Cz/CzechAnalyzer.cs
index 7a1c839..3807379 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Cz/CzechAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Cz/CzechAnalyzer.cs
@@ -75,8 +75,10 @@ namespace Lucene.Net.Analysis.Cz
                 {
                     DEFAULT_SET = WordlistLoader.GetWordSet(
                         IOUtils.GetDecodingReader(typeof(CzechAnalyzer), typeof(CzechAnalyzer).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8), 
-                        "#", 
+                        "#",
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {
@@ -145,7 +147,9 @@ namespace Lucene.Net.Analysis.Cz
             TokenStream result = new StandardFilter(matchVersion, source);
             result = new LowerCaseFilter(matchVersion, result);
             result = new StopFilter(matchVersion, result, stopwords);
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 if (this.stemExclusionTable.Any())
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Da/DanishAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Da/DanishAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Da/DanishAnalyzer.cs
index f9142bd..74ee9f6 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Da/DanishAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Da/DanishAnalyzer.cs
@@ -64,7 +64,9 @@ namespace Lucene.Net.Analysis.Da
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
                         IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/De/GermanAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/De/GermanAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/De/GermanAnalyzer.cs
index 6f79f93..ff0e204 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/De/GermanAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/De/GermanAnalyzer.cs
@@ -91,7 +91,9 @@ namespace Lucene.Net.Analysis.De
                 {
                     DEFAULT_SET = WordlistLoader.GetSnowballWordSet(
                         IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {
@@ -116,7 +118,10 @@ namespace Lucene.Net.Analysis.De
         /// <seealso cref="#getDefaultStopSet()"/>.
         /// </summary>
         public GermanAnalyzer(LuceneVersion matchVersion)
-              : this(matchVersion, matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ? DefaultSetHolder.DEFAULT_SET : DefaultSetHolder.DEFAULT_SET_30)
+#pragma warning disable 612, 618
+              : this(matchVersion, matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ? 
+                    DefaultSetHolder.DEFAULT_SET : DefaultSetHolder.DEFAULT_SET_30)
+#pragma warning restore 612, 618
         {
         }
 
@@ -164,12 +169,14 @@ namespace Lucene.Net.Analysis.De
             result = new LowerCaseFilter(matchVersion, result);
             result = new StopFilter(matchVersion, result, stopwords);
             result = new SetKeywordMarkerFilter(result, exclusionSet);
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
             {
                 result = new GermanNormalizationFilter(result);
                 result = new GermanLightStemFilter(result);
             }
             else if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 result = new SnowballFilter(result, new German2Stemmer());
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/El/GreekAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/El/GreekAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/El/GreekAnalyzer.cs
index fad5003..573a8ed 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/El/GreekAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/El/GreekAnalyzer.cs
@@ -119,12 +119,14 @@ namespace Lucene.Net.Analysis.El
         {
             Tokenizer source = new StandardTokenizer(matchVersion, reader);
             TokenStream result = new GreekLowerCaseFilter(matchVersion, source);
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
             {
                 result = new StandardFilter(matchVersion, result);
             }
             result = new StopFilter(matchVersion, result, stopwords);
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 result = new GreekStemFilter(result);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/El/GreekStemmer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/El/GreekStemmer.cs b/src/Lucene.Net.Analysis.Common/Analysis/El/GreekStemmer.cs
index 8927d94..dbf3289 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/El/GreekStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/El/GreekStemmer.cs
@@ -256,7 +256,9 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
+#pragma warning disable 612, 618
         private static readonly CharArraySet exc4 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, Arrays.AsList("\u03b8", "\u03b4", "\u03b5\u03bb", "\u03b3\u03b1\u03bb", "\u03bd", "\u03c0", "\u03b9\u03b4", "\u03c0\u03b1\u03c1"), false);
+#pragma warning restore 612, 618
 
         private int Rule4(char[] s, int len)
         {
@@ -294,7 +296,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc6 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc6 =
+#pragma warning disable 612, 618
+            new CharArraySet(LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b1\u03bb", "\u03b1\u03b4", "\u03b5\u03bd\u03b4", "\u03b1\u03bc\u03b1\u03bd", "\u03b1\u03bc\u03bc\u03bf\u03c7\u03b1\u03bb", "\u03b7\u03b8", "\u03b1\u03bd\u03b7\u03b8", 
                 "\u03b1\u03bd\u03c4\u03b9\u03b4", "\u03c6\u03c5\u03c3", "\u03b2\u03c1\u03c9\u03bc", "\u03b3\u03b5\u03c1", "\u03b5\u03be\u03c9\u03b4", "\u03ba\u03b1\u03bb\u03c0", "\u03ba\u03b1\u03bb\u03bb\u03b9\u03bd", "\u03ba\u03b1\u03c4\u03b1\u03b4", 
                 "\u03bc\u03bf\u03c5\u03bb", "\u03bc\u03c0\u03b1\u03bd", "\u03bc\u03c0\u03b1\u03b3\u03b9\u03b1\u03c4", "\u03bc\u03c0\u03bf\u03bb", "\u03bc\u03c0\u03bf\u03c3", "\u03bd\u03b9\u03c4", "\u03be\u03b9\u03ba", "\u03c3\u03c5\u03bd\u03bf\u03bc\u03b7\u03bb", 
@@ -327,7 +332,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc7 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc7 =
+#pragma warning disable 612, 618
+            new CharArraySet(LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b1\u03bd\u03b1\u03c0", "\u03b1\u03c0\u03bf\u03b8", "\u03b1\u03c0\u03bf\u03ba", "\u03b1\u03c0\u03bf\u03c3\u03c4", "\u03b2\u03bf\u03c5\u03b2", "\u03be\u03b5\u03b8", "\u03bf\u03c5\u03bb", 
                 "\u03c0\u03b5\u03b8", "\u03c0\u03b9\u03ba\u03c1", "\u03c0\u03bf\u03c4", "\u03c3\u03b9\u03c7", "\u03c7"), false);
 
@@ -365,10 +373,16 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc8a = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc8a = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03c4\u03c1", "\u03c4\u03c3"), false);
 
-        private static readonly CharArraySet exc8b = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc8b = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b2\u03b5\u03c4\u03b5\u03c1", "\u03b2\u03bf\u03c5\u03bb\u03ba", "\u03b2\u03c1\u03b1\u03c7\u03bc", "\u03b3", "\u03b4\u03c1\u03b1\u03b4\u03bf\u03c5\u03bc", "\u03b8", "\u03ba\u03b1\u03bb\u03c0\u03bf\u03c5\u03b6", 
                 "\u03ba\u03b1\u03c3\u03c4\u03b5\u03bb", "\u03ba\u03bf\u03c1\u03bc\u03bf\u03c1", "\u03bb\u03b1\u03bf\u03c0\u03bb", "\u03bc\u03c9\u03b1\u03bc\u03b5\u03b8", "\u03bc", "\u03bc\u03bf\u03c5\u03c3\u03bf\u03c5\u03bb\u03bc", "\u03bd", "\u03bf\u03c5\u03bb", 
                 "\u03c0", "\u03c0\u03b5\u03bb\u03b5\u03ba", "\u03c0\u03bb", "\u03c0\u03bf\u03bb\u03b9\u03c3", "\u03c0\u03bf\u03c1\u03c4\u03bf\u03bb", "\u03c3\u03b1\u03c1\u03b1\u03ba\u03b1\u03c4\u03c3", "\u03c3\u03bf\u03c5\u03bb\u03c4", 
@@ -437,7 +451,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc9 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc9 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b1\u03b2\u03b1\u03c1", "\u03b2\u03b5\u03bd", "\u03b5\u03bd\u03b1\u03c1", "\u03b1\u03b2\u03c1", "\u03b1\u03b4", "\u03b1\u03b8", "\u03b1\u03bd", "\u03b1\u03c0\u03bb", 
                 "\u03b2\u03b1\u03c1\u03bf\u03bd", "\u03bd\u03c4\u03c1", "\u03c3\u03ba", "\u03ba\u03bf\u03c0", "\u03bc\u03c0\u03bf\u03c1", "\u03bd\u03b9\u03c6", "\u03c0\u03b1\u03b3", "\u03c0\u03b1\u03c1\u03b1\u03ba\u03b1\u03bb", "\u03c3\u03b5\u03c1\u03c0", 
                 "\u03c3\u03ba\u03b5\u03bb", "\u03c3\u03c5\u03c1\u03c6", "\u03c4\u03bf\u03ba", "\u03c5", "\u03b4", "\u03b5\u03bc", "\u03b8\u03b1\u03c1\u03c1", "\u03b8"), false);
@@ -539,11 +556,13 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
+#pragma warning disable 612, 618
         private static readonly CharArraySet exc12a = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
             Arrays.AsList("\u03c0", "\u03b1\u03c0", "\u03c3\u03c5\u03bc\u03c0", "\u03b1\u03c3\u03c5\u03bc\u03c0", "\u03b1\u03ba\u03b1\u03c4\u03b1\u03c0", "\u03b1\u03bc\u03b5\u03c4\u03b1\u03bc\u03c6"), false);
 
         private static readonly CharArraySet exc12b = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
             Arrays.AsList("\u03b1\u03bb", "\u03b1\u03c1", "\u03b5\u03ba\u03c4\u03b5\u03bb", "\u03b6", "\u03bc", "\u03be", "\u03c0\u03b1\u03c1\u03b1\u03ba\u03b1\u03bb", "\u03b1\u03c1", "\u03c0\u03c1\u03bf", "\u03bd\u03b9\u03c3"), false);
+#pragma warning restore 612, 618
 
         private int Rule12(char[] s, int len)
         {
@@ -568,7 +587,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc13 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc13 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b4\u03b9\u03b1\u03b8", "\u03b8", "\u03c0\u03b1\u03c1\u03b1\u03ba\u03b1\u03c4\u03b1\u03b8", "\u03c0\u03c1\u03bf\u03c3\u03b8", "\u03c3\u03c5\u03bd\u03b8"), false);
 
         private int Rule13(char[] s, int len)
@@ -609,7 +631,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc14 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc14 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03c6\u03b1\u03c1\u03bc\u03b1\u03ba", "\u03c7\u03b1\u03b4", "\u03b1\u03b3\u03ba", "\u03b1\u03bd\u03b1\u03c1\u03c1", "\u03b2\u03c1\u03bf\u03bc", "\u03b5\u03ba\u03bb\u03b9\u03c0", "\u03bb\u03b1\u03bc\u03c0\u03b9\u03b4", 
                 "\u03bb\u03b5\u03c7", "\u03bc", "\u03c0\u03b1\u03c4", "\u03c1", "\u03bb", "\u03bc\u03b5\u03b4", "\u03bc\u03b5\u03c3\u03b1\u03b6", "\u03c5\u03c0\u03bf\u03c4\u03b5\u03b9\u03bd", "\u03b1\u03bc", "\u03b1\u03b9\u03b8", 
                 "\u03b1\u03bd\u03b7\u03ba", "\u03b4\u03b5\u03c3\u03c0\u03bf\u03b6", "\u03b5\u03bd\u03b4\u03b9\u03b1\u03c6\u03b5\u03c1", "\u03b4\u03b5", "\u03b4\u03b5\u03c5\u03c4\u03b5\u03c1\u03b5\u03c5", "\u03ba\u03b1\u03b8\u03b1\u03c1\u03b5\u03c5", "\u03c0\u03bb\u03b5", "\u03c4\u03c3\u03b1"), false);
@@ -651,7 +676,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc15a = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc15a = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b1\u03b2\u03b1\u03c3\u03c4", "\u03c0\u03bf\u03bb\u03c5\u03c6", "\u03b1\u03b4\u03b7\u03c6", "\u03c0\u03b1\u03bc\u03c6", "\u03c1", "\u03b1\u03c3\u03c0", "\u03b1\u03c6", "\u03b1\u03bc\u03b1\u03bb", 
                 "\u03b1\u03bc\u03b1\u03bb\u03bb\u03b9", "\u03b1\u03bd\u03c5\u03c3\u03c4", "\u03b1\u03c0\u03b5\u03c1", "\u03b1\u03c3\u03c0\u03b1\u03c1", "\u03b1\u03c7\u03b1\u03c1", "\u03b4\u03b5\u03c1\u03b2\u03b5\u03bd", "\u03b4\u03c1\u03bf\u03c3\u03bf\u03c0", 
                 "\u03be\u03b5\u03c6", "\u03bd\u03b5\u03bf\u03c0", "\u03bd\u03bf\u03bc\u03bf\u03c4", "\u03bf\u03bb\u03bf\u03c0", "\u03bf\u03bc\u03bf\u03c4", "\u03c0\u03c1\u03bf\u03c3\u03c4", "\u03c0\u03c1\u03bf\u03c3\u03c9\u03c0\u03bf\u03c0", "\u03c3\u03c5\u03bc\u03c0", 
@@ -659,7 +687,11 @@ namespace Lucene.Net.Analysis.El
                 "\u03b1\u03c1\u03c4\u03b9\u03c0", "\u03b4\u03b9\u03b1\u03c4", "\u03b5\u03bd", "\u03b5\u03c0\u03b9\u03c4", "\u03ba\u03c1\u03bf\u03ba\u03b1\u03bb\u03bf\u03c0", "\u03c3\u03b9\u03b4\u03b7\u03c1\u03bf\u03c0", "\u03bb", "\u03bd\u03b1\u03c5", 
                 "\u03bf\u03c5\u03bb\u03b1\u03bc", "\u03bf\u03c5\u03c1", "\u03c0", "\u03c4\u03c1", "\u03bc"), false);
 
-        private static readonly CharArraySet exc15b = new CharArraySet(LuceneVersion.LUCENE_CURRENT, Arrays.AsList("\u03c8\u03bf\u03c6", "\u03bd\u03b1\u03c5\u03bb\u03bf\u03c7"), false);
+        private static readonly CharArraySet exc15b = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
+            Arrays.AsList("\u03c8\u03bf\u03c6", "\u03bd\u03b1\u03c5\u03bb\u03bf\u03c7"), false);
 
         private int Rule15(char[] s, int len)
         {
@@ -700,7 +732,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc16 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc16 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03bd", "\u03c7\u03b5\u03c1\u03c3\u03bf\u03bd", "\u03b4\u03c9\u03b4\u03b5\u03ba\u03b1\u03bd", "\u03b5\u03c1\u03b7\u03bc\u03bf\u03bd", "\u03bc\u03b5\u03b3\u03b1\u03bb\u03bf\u03bd", "\u03b5\u03c0\u03c4\u03b1\u03bd"), false);
 
         private int Rule16(char[] s, int len)
@@ -725,7 +760,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc17 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc17 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03b1\u03c3\u03b2", "\u03c3\u03b2", "\u03b1\u03c7\u03c1", "\u03c7\u03c1", "\u03b1\u03c0\u03bb", "\u03b1\u03b5\u03b9\u03bc\u03bd", "\u03b4\u03c5\u03c3\u03c7\u03c1", "\u03b5\u03c5\u03c7\u03c1", "\u03ba\u03bf\u03b9\u03bd\u03bf\u03c7\u03c1", "\u03c0\u03b1\u03bb\u03b9\u03bc\u03c8"), false);
 
         private int Rule17(char[] s, int len)
@@ -742,7 +780,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc18 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc18 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03bd", "\u03c1", "\u03c3\u03c0\u03b9", "\u03c3\u03c4\u03c1\u03b1\u03b2\u03bf\u03bc\u03bf\u03c5\u03c4\u03c3", "\u03ba\u03b1\u03ba\u03bf\u03bc\u03bf\u03c5\u03c4\u03c3", "\u03b5\u03be\u03c9\u03bd"), false);
 
         private int Rule18(char[] s, int len)
@@ -770,7 +811,10 @@ namespace Lucene.Net.Analysis.El
             return len;
         }
 
-        private static readonly CharArraySet exc19 = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 
+        private static readonly CharArraySet exc19 = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             Arrays.AsList("\u03c0\u03b1\u03c1\u03b1\u03c3\u03bf\u03c5\u03c3", "\u03c6", "\u03c7", "\u03c9\u03c1\u03b9\u03bf\u03c0\u03bb", "\u03b1\u03b6", "\u03b1\u03bb\u03bb\u03bf\u03c3\u03bf\u03c5\u03c3", "\u03b1\u03c3\u03bf\u03c5\u03c3"), false);
 
         private int Rule19(char[] s, int len)

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishAnalyzer.cs
index 381c01d..800cd46 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishAnalyzer.cs
@@ -101,7 +101,9 @@ namespace Lucene.Net.Analysis.En
             Tokenizer source = new StandardTokenizer(matchVersion, reader);
             TokenStream result = new StandardFilter(matchVersion, source);
             // prior to this we get the classic behavior, standardfilter does it for us.
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 result = new EnglishPossessiveFilter(matchVersion, result);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishPossessiveFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishPossessiveFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishPossessiveFilter.cs
index 4e25ac4..a4715e7 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishPossessiveFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/En/EnglishPossessiveFilter.cs
@@ -59,7 +59,11 @@ namespace Lucene.Net.Analysis.En
             char[] buffer = termAtt.Buffer();
             int bufferLength = termAtt.Length;
 
-            if (bufferLength >= 2 && (buffer[bufferLength - 2] == '\'' || (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36) && (buffer[bufferLength - 2] == '\u2019' || buffer[bufferLength - 2] == '\uFF07'))) && (buffer[bufferLength - 1] == 's' || buffer[bufferLength - 1] == 'S'))
+            if (bufferLength >= 2 && (buffer[bufferLength - 2] == '\'' ||
+#pragma warning disable 612, 618
+                (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36) && (buffer[bufferLength - 2] == '\u2019' ||
+#pragma warning restore 612, 618
+                buffer[bufferLength - 2] == '\uFF07'))) && (buffer[bufferLength - 1] == 's' || buffer[bufferLength - 1] == 'S'))
             {
                 termAtt.Length = bufferLength - 2; // Strip last 2 characters off
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/En/KStemmer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/En/KStemmer.cs b/src/Lucene.Net.Analysis.Common/Analysis/En/KStemmer.cs
index fefda71..d04c64f 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/En/KStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/En/KStemmer.cs
@@ -408,7 +408,9 @@ namespace Lucene.Net.Analysis.En
             DictEntry defaultEntry;
             DictEntry entry;
 
+#pragma warning disable 612, 618
             CharArrayMap<DictEntry> d = new CharArrayMap<DictEntry>(LuceneVersion.LUCENE_CURRENT, 1000, false);
+#pragma warning restore 612, 618
             for (int i = 0; i < exceptionWords.Length; i++)
             {
                 if (!d.ContainsKey(exceptionWords[i]))

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Es/SpanishAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Es/SpanishAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Es/SpanishAnalyzer.cs
index d55e459..1cb6d31 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Es/SpanishAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Es/SpanishAnalyzer.cs
@@ -73,7 +73,9 @@ namespace Lucene.Net.Analysis.Es
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
                         IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {
@@ -137,7 +139,9 @@ namespace Lucene.Net.Analysis.Es
             {
                 result = new SetKeywordMarkerFilter(result, stemExclusionSet);
             }
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
+#pragma warning restore 612, 618
             {
                 result = new SpanishLightStemFilter(result);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Fa/PersianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Fa/PersianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Fa/PersianAnalyzer.cs
index 59855c1..9aa4a6f 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Fa/PersianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Fa/PersianAnalyzer.cs
@@ -120,13 +120,17 @@ namespace Lucene.Net.Analysis.Fa
         public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             Tokenizer source;
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 source = new StandardTokenizer(matchVersion, reader);
             }
             else
             {
+#pragma warning disable 612, 618
                 source = new ArabicLetterTokenizer(matchVersion, reader);
+#pragma warning restore 612, 618
             }
             TokenStream result = new LowerCaseFilter(matchVersion, source);
             result = new ArabicNormalizationFilter(result);
@@ -144,7 +148,10 @@ namespace Lucene.Net.Analysis.Fa
         /// </summary>
         public override TextReader InitReader(string fieldName, TextReader reader)
         {
-            return matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ? new PersianCharFilter(reader) : reader;
+#pragma warning disable 612, 618
+            return matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ?
+#pragma warning restore 612, 618
+                new PersianCharFilter(reader) : reader;
         }
     }
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Fi/FinnishAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Fi/FinnishAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Fi/FinnishAnalyzer.cs
index 03a786c..3ca82a6 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Fi/FinnishAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Fi/FinnishAnalyzer.cs
@@ -64,7 +64,9 @@ namespace Lucene.Net.Analysis.Fi
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
                         IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Fr/FrenchAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Fr/FrenchAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Fr/FrenchAnalyzer.cs
index af22869..9768cfe 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Fr/FrenchAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Fr/FrenchAnalyzer.cs
@@ -69,7 +69,10 @@ namespace Lucene.Net.Analysis.Fr
 
         /// <summary>
         /// Default set of articles for ElisionFilter </summary>
-        public static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(new CharArraySet(LuceneVersion.LUCENE_CURRENT,
+        public static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
             new string[] { "l", "m", "t", "qu", "n", "s", "j", "d", "c", "jusqu", "quoiqu", "lorsqu", "puisqu" }, true));
 
         /// <summary>
@@ -99,8 +102,10 @@ namespace Lucene.Net.Analysis.Fr
                 try
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
-                        IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8), 
+                        IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {
@@ -115,7 +120,10 @@ namespace Lucene.Net.Analysis.Fr
         /// Builds an analyzer with the default stop words (<seealso cref="#getDefaultStopSet"/>).
         /// </summary>
         public FrenchAnalyzer(LuceneVersion matchVersion)
-              : this(matchVersion, matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ? DefaultSetHolder.DEFAULT_STOP_SET : DefaultSetHolder.DEFAULT_STOP_SET_30)
+#pragma warning disable 612, 618
+              : this(matchVersion, matchVersion.OnOrAfter(LuceneVersion.LUCENE_31) ? 
+                    DefaultSetHolder.DEFAULT_STOP_SET : DefaultSetHolder.DEFAULT_STOP_SET_30)
+#pragma warning restore 612, 618
         {
         }
 
@@ -160,7 +168,9 @@ namespace Lucene.Net.Analysis.Fr
         ///         
         public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 Tokenizer source = new StandardTokenizer(matchVersion, reader);
                 TokenStream result = new StandardFilter(matchVersion, source);
@@ -171,7 +181,9 @@ namespace Lucene.Net.Analysis.Fr
                 {
                     result = new SetKeywordMarkerFilter(result, excltable);
                 }
+#pragma warning disable 612, 618
                 if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
+#pragma warning restore 612, 618
                 {
                     result = new FrenchLightStemFilter(result);
                 }
@@ -190,7 +202,9 @@ namespace Lucene.Net.Analysis.Fr
                 {
                     result = new SetKeywordMarkerFilter(result, excltable);
                 }
+#pragma warning disable 612, 618
                 result = new FrenchStemFilter(result);
+#pragma warning restore 612, 618
                 // Convert to lowercase after stemming!
                 return new TokenStreamComponents(source, new LowerCaseFilter(matchVersion, result));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ga/IrishAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ga/IrishAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ga/IrishAnalyzer.cs
index a2bcb6a..47b1a67 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ga/IrishAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ga/IrishAnalyzer.cs
@@ -39,14 +39,22 @@ namespace Lucene.Net.Analysis.Ga
         /// File containing default Irish stopwords. </summary>
         public const string DEFAULT_STOPWORD_FILE = "stopwords.txt";
 
-        private static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(new CharArraySet(LuceneVersion.LUCENE_CURRENT, Arrays.AsList("d", "m", "b"), true));
+        private static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
+            Arrays.AsList("d", "m", "b"), true));
 
         /// <summary>
         /// When StandardTokenizer splits t\u2011athair into {t, athair}, we don't
         /// want to cause a position increment, otherwise there will be problems
         /// with phrase queries versus tAthair (which would not have a gap).
         /// </summary>
-        private static readonly CharArraySet HYPHENATIONS = CharArraySet.UnmodifiableSet(new CharArraySet(LuceneVersion.LUCENE_CURRENT, Arrays.AsList("h", "n", "t"), true));
+        private static readonly CharArraySet HYPHENATIONS = CharArraySet.UnmodifiableSet(new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
+            Arrays.AsList("h", "n", "t"), true));
 
         /// <summary>
         /// Returns an unmodifiable instance of the default stop words set. </summary>
@@ -130,7 +138,9 @@ namespace Lucene.Net.Analysis.Ga
             Tokenizer source = new StandardTokenizer(matchVersion, reader);
             TokenStream result = new StandardFilter(matchVersion, source);
             StopFilter s = new StopFilter(matchVersion, result, HYPHENATIONS);
+#pragma warning disable 612, 618
             if (!matchVersion.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
             {
                 s.EnablePositionIncrements = false;
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Gl/GalicianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Gl/GalicianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Gl/GalicianAnalyzer.cs
index c8b4afa..c4db0fc 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Gl/GalicianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Gl/GalicianAnalyzer.cs
@@ -62,7 +62,9 @@ namespace Lucene.Net.Analysis.Gl
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetWordSet(
                         IOUtils.GetDecodingReader(typeof(GalicianAnalyzer), typeof(GalicianAnalyzer).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException ex)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Hi/HindiAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Hi/HindiAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Hi/HindiAnalyzer.cs
index f458bd5..654c747 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Hi/HindiAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Hi/HindiAnalyzer.cs
@@ -130,6 +130,7 @@ namespace Lucene.Net.Analysis.Hi
         public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             Tokenizer source;
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
             {
                 source = new StandardTokenizer(matchVersion, reader);
@@ -138,6 +139,7 @@ namespace Lucene.Net.Analysis.Hi
             {
                 source = new IndicTokenizer(matchVersion, reader);
             }
+#pragma warning restore 612, 618
             TokenStream result = new LowerCaseFilter(matchVersion, source);
             if (stemExclusionSet.Count > 0)
             {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Hu/HungarianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Hu/HungarianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Hu/HungarianAnalyzer.cs
index 95a79a8..00b4d84 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Hu/HungarianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Hu/HungarianAnalyzer.cs
@@ -63,8 +63,10 @@ namespace Lucene.Net.Analysis.Hu
                 try
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
-                        IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8), 
+                        IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Stemmer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Stemmer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Stemmer.cs
index 8e0070d..360fe4f 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Stemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Stemmer.cs
@@ -109,7 +109,10 @@ namespace Lucene.Net.Analysis.Hunspell
             {
                 return stems;
             }
-            CharArraySet terms = new CharArraySet(LuceneVersion.LUCENE_CURRENT, 8, dictionary.ignoreCase);
+            CharArraySet terms = new CharArraySet(
+#pragma warning disable 612, 618
+                LuceneVersion.LUCENE_CURRENT, 8, dictionary.ignoreCase);
+#pragma warning restore 612, 618
             IList<CharsRef> deduped = new List<CharsRef>();
             foreach (CharsRef s in stems)
             {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/It/ItalianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/It/ItalianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/It/ItalianAnalyzer.cs
index 44b8449..1ff4426 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/It/ItalianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/It/ItalianAnalyzer.cs
@@ -51,8 +51,12 @@ namespace Lucene.Net.Analysis.It
         /// File containing default Italian stopwords. </summary>
         public const string DEFAULT_STOPWORD_FILE = "italian_stop.txt";
 
-        private static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(new CharArraySet(LuceneVersion.LUCENE_CURRENT,
-            Arrays.AsList("c", "l", "all", "dall", "dell", "nell", "sull", "coll", "pell", "gl", "agl", "dagl", "degl", "negl", "sugl", "un", "m", "t", "s", "v", "d"), true));
+        private static readonly CharArraySet DEFAULT_ARTICLES = CharArraySet.UnmodifiableSet(new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_CURRENT,
+#pragma warning restore 612, 618
+            Arrays.AsList("c", "l", "all", "dall", "dell", "nell", "sull", "coll", "pell", "gl", "agl", 
+                "dagl", "degl", "negl", "sugl", "un", "m", "t", "s", "v", "d"), true));
 
         /// <summary>
         /// Returns an unmodifiable instance of the default stop words set. </summary>
@@ -79,7 +83,9 @@ namespace Lucene.Net.Analysis.It
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
                         IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {
@@ -137,7 +143,9 @@ namespace Lucene.Net.Analysis.It
         {
             Tokenizer source = new StandardTokenizer(matchVersion, reader);
             TokenStream result = new StandardFilter(matchVersion, source);
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_32))
+#pragma warning restore 612, 618
             {
                 result = new ElisionFilter(result, DEFAULT_ARTICLES);
             }
@@ -147,7 +155,9 @@ namespace Lucene.Net.Analysis.It
             {
                 result = new SetKeywordMarkerFilter(result, stemExclusionSet);
             }
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_36))
+#pragma warning restore 612, 618
             {
                 result = new ItalianLightStemFilter(result);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Lv/LatvianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Lv/LatvianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Lv/LatvianAnalyzer.cs
index be89de0..d0a4e98 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Lv/LatvianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Lv/LatvianAnalyzer.cs
@@ -61,8 +61,10 @@ namespace Lucene.Net.Analysis.Lv
                 try
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetWordSet(
-                        IOUtils.GetDecodingReader(typeof(LatvianAnalyzer), typeof(LatvianAnalyzer).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8), 
+                        IOUtils.GetDecodingReader(typeof(LatvianAnalyzer), typeof(LatvianAnalyzer).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/KeepWordFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/KeepWordFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/KeepWordFilterFactory.cs
index e120a60..1972e8e 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/KeepWordFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/KeepWordFilterFactory.cs
@@ -93,7 +93,9 @@ namespace Lucene.Net.Analysis.Miscellaneous
             }
             else
             {
+#pragma warning disable 612, 618
                 TokenStream filter = new KeepWordFilter(luceneMatchVersion, enablePositionIncrements, input, words);
+#pragma warning restore 612, 618
                 return filter;
             }
         }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/LengthFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/LengthFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/LengthFilterFactory.cs
index e462eb4..971c259 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/LengthFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/LengthFilterFactory.cs
@@ -54,7 +54,9 @@ namespace Lucene.Net.Analysis.Miscellaneous
 
         public override TokenStream Create(TokenStream input)
         {
+#pragma warning disable 612, 618
             var filter = new LengthFilter(luceneMatchVersion, enablePositionIncrements, input, min, max);
+#pragma warning restore 612, 618
             return filter;
         }
     }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/RemoveDuplicatesTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/RemoveDuplicatesTokenFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/RemoveDuplicatesTokenFilter.cs
index 88e0602..ec3386b 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/RemoveDuplicatesTokenFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/RemoveDuplicatesTokenFilter.cs
@@ -32,7 +32,10 @@ namespace Lucene.Net.Analysis.Miscellaneous
         private readonly IPositionIncrementAttribute posIncAttribute;
 
         // use a fixed version, as we don't care about case sensitivity.
-        private readonly CharArraySet previous = new CharArraySet(LuceneVersion.LUCENE_31, 8, false);
+        private readonly CharArraySet previous = new CharArraySet(
+#pragma warning disable 612, 618
+            LuceneVersion.LUCENE_31, 8, false);
+#pragma warning restore 612, 618
 
         /// <summary>
         /// Creates a new RemoveDuplicatesTokenFilter

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilter.cs
index e49e4e8..193ece1 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilter.cs
@@ -55,7 +55,9 @@ namespace Lucene.Net.Analysis.Miscellaneous
         /// <summary>
         /// Create a new <seealso cref="TrimFilter"/> on top of <code>in</code>. </summary>
         public TrimFilter(LuceneVersion version, TokenStream @in)
+#pragma warning disable 612, 618
             : this(version, @in, false)
+#pragma warning restore 612, 618
         {
         }
 

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilterFactory.cs
index 0c01908..297d0b5 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/TrimFilterFactory.cs
@@ -50,7 +50,9 @@ namespace Lucene.Net.Analysis.Miscellaneous
 
         public override TokenStream Create(TokenStream input)
         {
+#pragma warning disable 612, 618
             var filter = new TrimFilter(luceneMatchVersion, input, updateOffsets);
+#pragma warning restore 612, 618
             return filter;
         }
     }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/WordDelimiterFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/WordDelimiterFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/WordDelimiterFilterFactory.cs
index 25258ce..00671b1 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/WordDelimiterFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Miscellaneous/WordDelimiterFilterFactory.cs
@@ -130,7 +130,10 @@ namespace Lucene.Net.Analysis.Miscellaneous
             }
             else
             {
-                return new Lucene47WordDelimiterFilter(input, typeTable ?? WordDelimiterIterator.DEFAULT_WORD_DELIM_TABLE, flags, protectedWords);
+#pragma warning disable 612, 618
+                return new Lucene47WordDelimiterFilter(
+#pragma warning restore 612, 618
+                    input, typeTable ?? WordDelimiterIterator.DEFAULT_WORD_DELIM_TABLE, flags, protectedWords);
             }
         }
 

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramFilterFactory.cs
index 0eef8fc..3c6a3c1 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramFilterFactory.cs
@@ -52,7 +52,9 @@ namespace Lucene.Net.Analysis.Ngram
 
         public override TokenStream Create(TokenStream input)
         {
+#pragma warning disable 612, 618
             return new EdgeNGramTokenFilter(luceneMatchVersion, input, side, minGramSize, maxGramSize);
+#pragma warning restore 612, 618
         }
     }
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenFilter.cs
index 76400bf..5bb5d15 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenFilter.cs
@@ -98,17 +98,17 @@ namespace Lucene.Net.Analysis.Ngram
               : base(input)
         {
 
-            if (version == null)
-            {
-                throw new System.ArgumentException("version must not be null");
-            }
+            //if (version == null)
+            //{
+            //    throw new System.ArgumentException("version must not be null");
+            //}
 
             if (version.OnOrAfter(LuceneVersion.LUCENE_44) && side == Side.BACK)
             {
                 throw new System.ArgumentException("Side.BACK is not supported anymore as of Lucene 4.4, use ReverseStringFilter up-front and afterward");
             }
 
-            if (side == null)
+            if (!Enum.IsDefined(typeof(Side), side))
             {
                 throw new System.ArgumentException("sideLabel must be either front or back");
             }
@@ -157,7 +157,9 @@ namespace Lucene.Net.Analysis.Ngram
         /// <param name="minGram"> the smallest n-gram to generate </param>
         /// <param name="maxGram"> the largest n-gram to generate </param>
         public EdgeNGramTokenFilter(LuceneVersion version, TokenStream input, int minGram, int maxGram)
+#pragma warning disable 612, 618
               : this(version, input, Side.FRONT, minGram, maxGram)
+#pragma warning restore 612, 618
         {
         }
 
@@ -179,7 +181,9 @@ namespace Lucene.Net.Analysis.Ngram
                         curGramSize = minGram;
                         tokStart = offsetAtt.StartOffset();
                         tokEnd = offsetAtt.EndOffset();
+#pragma warning disable 612, 618
                         if (version.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
                         {
                             // Never update offsets
                             updateOffsets = false;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenizerFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenizerFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenizerFactory.cs
index 7f7b645..dbf0e12 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenizerFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/EdgeNGramTokenizerFactory.cs
@@ -53,7 +53,9 @@ namespace Lucene.Net.Analysis.Ngram
 
         public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
         {
+#pragma warning disable 612, 618
             if (luceneMatchVersion.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
             {
                 EdgeNGramTokenFilter.Side sideEnum;
                 if (!Enum.TryParse(this.side, true, out sideEnum))
@@ -64,7 +66,9 @@ namespace Lucene.Net.Analysis.Ngram
             }
             else
             {
+#pragma warning disable 612, 618
                 return new Lucene43EdgeNGramTokenizer(luceneMatchVersion, input, side, minGramSize, maxGramSize);
+#pragma warning restore 612, 618
             }
         }
     }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/Lucene43EdgeNGramTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/Lucene43EdgeNGramTokenizer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/Lucene43EdgeNGramTokenizer.cs
index b735b3d..96173c3 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/Lucene43EdgeNGramTokenizer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/Lucene43EdgeNGramTokenizer.cs
@@ -159,12 +159,12 @@ namespace Lucene.Net.Analysis.Ngram
 
         private void Init(LuceneVersion version, Side side, int minGram, int maxGram)
         {
-            if (version == null)
-            {
-                throw new System.ArgumentException("version must not be null");
-            }
+            //if (version == null)
+            //{
+            //    throw new System.ArgumentException("version must not be null");
+            //}
 
-            if (side == null)
+            if (!Enum.IsDefined(typeof(Side), side))
             {
                 throw new System.ArgumentException("sideLabel must be either front or back");
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenFilter.cs
index cdaa388..4ea8944 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenFilter.cs
@@ -80,7 +80,11 @@ namespace Lucene.Net.Analysis.Ngram
             : base(new CodepointCountFilter(version, input, minGram, int.MaxValue))
         {
             this.version = version;
-            this.charUtils = version.OnOrAfter(LuceneVersion.LUCENE_44) ? CharacterUtils.GetInstance(version) : CharacterUtils.Java4Instance;
+            this.charUtils = version.OnOrAfter(
+#pragma warning disable 612, 618
+                LuceneVersion.LUCENE_44) ?
+#pragma warning restore 612, 618
+                CharacterUtils.GetInstance(version) : CharacterUtils.Java4Instance;
             if (minGram < 1)
             {
                 throw new System.ArgumentException("minGram must be greater than zero");
@@ -91,7 +95,9 @@ namespace Lucene.Net.Analysis.Ngram
             }
             this.minGram = minGram;
             this.maxGram = maxGram;
+#pragma warning disable 612, 618
             if (version.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
             {
                 posIncAtt = AddAttribute<IPositionIncrementAttribute>();
                 posLenAtt = AddAttribute<IPositionLengthAttribute>();
@@ -186,7 +192,9 @@ namespace Lucene.Net.Analysis.Ngram
                         hasIllegalOffsets = (tokStart + curTermLength) != tokEnd;
                     }
                 }
+#pragma warning disable 612, 618
                 if (version.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
                 {
                     if (curGramSize > maxGram || (curPos + curGramSize) > curCodePointCount)
                     {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizer.cs
index 994210b..0146574 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizer.cs
@@ -126,11 +126,16 @@ namespace Lucene.Net.Analysis.Ngram
 
         private void Init(LuceneVersion version, int minGram, int maxGram, bool edgesOnly)
         {
+#pragma warning disable 612, 618
             if (!version.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
             {
                 throw new System.ArgumentException("This class only works with Lucene 4.4+. To emulate the old (broken) behavior of NGramTokenizer, use Lucene43NGramTokenizer/Lucene43EdgeNGramTokenizer");
             }
-            charUtils = version.OnOrAfter(LuceneVersion.LUCENE_44) ? CharacterUtils.GetInstance(version) : CharacterUtils.Java4Instance;
+#pragma warning disable 612, 618
+            charUtils = version.OnOrAfter(LuceneVersion.LUCENE_44) ?
+#pragma warning restore 612, 618
+                CharacterUtils.GetInstance(version) : CharacterUtils.Java4Instance;
             if (minGram < 1)
             {
                 throw new System.ArgumentException("minGram must be greater than zero");

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizerFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizerFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizerFactory.cs
index 9f48512..81f5b6c 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizerFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ngram/NGramTokenizerFactory.cs
@@ -53,13 +53,17 @@ namespace Lucene.Net.Analysis.Ngram
         /// Creates the <seealso cref="TokenStream"/> of n-grams from the given <seealso cref="TextReader"/> and <seealso cref="AttributeSource.AttributeFactory"/>. </summary>
         public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
         {
+#pragma warning disable 612, 618
             if (luceneMatchVersion.OnOrAfter(LuceneVersion.LUCENE_44))
+#pragma warning restore 612, 618
             {
                 return new NGramTokenizer(luceneMatchVersion, factory, input, minGramSize, maxGramSize);
             }
             else
             {
+#pragma warning disable 612, 618
                 return new Lucene43NGramTokenizer(factory, input, minGramSize, maxGramSize);
+#pragma warning restore 612, 618
             }
         }
     }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Nl/DutchAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Nl/DutchAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Nl/DutchAnalyzer.cs
index cc9784f..1c54b05 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Nl/DutchAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Nl/DutchAnalyzer.cs
@@ -83,8 +83,10 @@ namespace Lucene.Net.Analysis.Nl
                 try
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
-                        IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8), 
+                        IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {
@@ -92,8 +94,9 @@ namespace Lucene.Net.Analysis.Nl
                     // distribution (JAR)
                     throw new Exception("Unable to load default stopword set");
                 }
-
+#pragma warning disable 612, 618
                 DEFAULT_STEM_DICT = new CharArrayMap<string>(LuceneVersion.LUCENE_CURRENT, 4, false);
+#pragma warning restore 612, 618
                 DEFAULT_STEM_DICT.Put("fiets", "fiets"); //otherwise fiet
                 DEFAULT_STEM_DICT.Put("bromfiets", "bromfiets"); //otherwise bromfiet
                 DEFAULT_STEM_DICT.Put("ei", "eier");
@@ -131,7 +134,10 @@ namespace Lucene.Net.Analysis.Nl
 
         public DutchAnalyzer(LuceneVersion matchVersion, CharArraySet stopwords)
               : this(matchVersion, stopwords, CharArraySet.EMPTY_SET,
-                    matchVersion.OnOrAfter(LuceneVersion.LUCENE_36) ? DefaultSetHolder.DEFAULT_STEM_DICT : CharArrayMap<string>.EmptyMap())
+#pragma warning disable 612, 618
+                    matchVersion.OnOrAfter(LuceneVersion.LUCENE_36) ?
+#pragma warning restore 612, 618
+                    DefaultSetHolder.DEFAULT_STEM_DICT : CharArrayMap<string>.EmptyMap())
         {
             // historically, this ctor never the stem dict!!!!!
             // so we populate it only for >= 3.6
@@ -139,7 +145,10 @@ namespace Lucene.Net.Analysis.Nl
 
         public DutchAnalyzer(LuceneVersion matchVersion, CharArraySet stopwords, CharArraySet stemExclusionTable)
               : this(matchVersion, stopwords, stemExclusionTable,
-                    matchVersion.OnOrAfter(LuceneVersion.LUCENE_36) ? DefaultSetHolder.DEFAULT_STEM_DICT : CharArrayMap<string>.EmptyMap())
+#pragma warning disable 612, 618
+                    matchVersion.OnOrAfter(LuceneVersion.LUCENE_36) ?
+#pragma warning restore 612, 618
+                    DefaultSetHolder.DEFAULT_STEM_DICT : CharArrayMap<string>.EmptyMap())
         {
             // historically, this ctor never the stem dict!!!!!
             // so we populate it only for >= 3.6
@@ -150,7 +159,9 @@ namespace Lucene.Net.Analysis.Nl
             this.matchVersion = matchVersion;
             this.stoptable = CharArraySet.UnmodifiableSet(CharArraySet.Copy(matchVersion, stopwords));
             this.excltable = CharArraySet.UnmodifiableSet(CharArraySet.Copy(matchVersion, stemExclusionTable));
+#pragma warning disable 612, 618
             if (stemOverrideDict.Count == 0 || !matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 this.stemdict = null;
                 this.origStemdict = CharArrayMap<string>.UnmodifiableMap(CharArrayMap<string>.Copy(matchVersion, stemOverrideDict));
@@ -189,7 +200,9 @@ namespace Lucene.Net.Analysis.Nl
         ///   <seealso cref="StemmerOverrideFilter"/>, and <seealso cref="SnowballFilter"/> </returns>
         public override TokenStreamComponents CreateComponents(string fieldName, TextReader aReader)
         {
+#pragma warning disable 612, 618
             if (matchVersion.OnOrAfter(LuceneVersion.LUCENE_31))
+#pragma warning restore 612, 618
             {
                 Tokenizer source = new StandardTokenizer(matchVersion, aReader);
                 TokenStream result = new StandardFilter(matchVersion, source);
@@ -215,7 +228,9 @@ namespace Lucene.Net.Analysis.Nl
                 {
                     result = new SetKeywordMarkerFilter(result, excltable);
                 }
+#pragma warning disable 612, 618
                 result = new DutchStemFilter(result, origStemdict);
+#pragma warning restore 612, 618
                 return new TokenStreamComponents(source, result);
             }
         }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/No/NorwegianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/No/NorwegianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/No/NorwegianAnalyzer.cs
index ddec099..0cb2dac 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/No/NorwegianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/No/NorwegianAnalyzer.cs
@@ -64,7 +64,9 @@ namespace Lucene.Net.Analysis.No
                 {
                     DEFAULT_STOP_SET = WordlistLoader.GetSnowballWordSet(
                         IOUtils.GetDecodingReader(typeof(SnowballFilter), typeof(SnowballFilter).Namespace + "." + DEFAULT_STOPWORD_FILE, Encoding.UTF8),
+#pragma warning disable 612, 618
                         LuceneVersion.LUCENE_CURRENT);
+#pragma warning restore 612, 618
                 }
                 catch (IOException)
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/da2ae4c8/src/Lucene.Net.Analysis.Common/Analysis/Pattern/PatternReplaceCharFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Pattern/PatternReplaceCharFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Pattern/PatternReplaceCharFilterFactory.cs
index 19e3c50..1d3a987 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Pattern/PatternReplaceCharFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Pattern/PatternReplaceCharFilterFactory.cs
@@ -49,7 +49,10 @@ namespace Lucene.Net.Analysis.Pattern
             pattern = GetPattern(args, "pattern");
             replacement = Get(args, "replacement", "");
             // TODO: warn if you set maxBlockChars or blockDelimiters ?
-            maxBlockChars = GetInt(args, "maxBlockChars", PatternReplaceCharFilter.DEFAULT_MAX_BLOCK_CHARS);
+            maxBlockChars = GetInt(args, "maxBlockChars",
+#pragma warning disable 612, 618
+                PatternReplaceCharFilter.DEFAULT_MAX_BLOCK_CHARS);
+#pragma warning restore 612, 618
             if (args.TryGetValue("blockDelimiters", out blockDelimiters))
             {
                 args.Remove("blockDelimiters");
@@ -62,7 +65,9 @@ namespace Lucene.Net.Analysis.Pattern
 
         public override TextReader Create(TextReader input)
         {
+#pragma warning disable 612, 618
             return new PatternReplaceCharFilter(pattern, replacement, maxBlockChars, blockDelimiters, input);
+#pragma warning restore 612, 618
         }
     }
 }
\ No newline at end of file