You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucenenet.apache.org by ni...@apache.org on 2022/10/31 00:33:56 UTC
[lucenenet] branch master updated: Sonar changes required for #671 (#730)
This is an automated email from the ASF dual-hosted git repository.
nightowl888 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/lucenenet.git
The following commit(s) were added to refs/heads/master by this push:
new 31ac9f6b7 Sonar changes required for #671 (#730)
31ac9f6b7 is described below
commit 31ac9f6b78a9ab75506d5f07da1f35a073de3c86
Author: Lakshay Sachdeva <35...@users.noreply.github.com>
AuthorDate: Mon Oct 31 06:03:49 2022 +0530
Sonar changes required for #671 (#730)
* Sonar changes required for #671
* Update QueryParser.cs
* changes related to review comments.
---
.../Compound/Hyphenation/HyphenationTree.cs | 2 +-
.../Analysis/Hunspell/Dictionary.cs | 2 +-
.../Analysis/Sinks/TeeSinkTokenFilter.cs | 2 +-
.../Tartarus/Snowball/Ext/BasqueStemmer.cs | 4 +-
.../Tartarus/Snowball/Ext/CatalanStemmer.cs | 2 +-
.../Tartarus/Snowball/Ext/EnglishStemmer.cs | 4 +-
.../Tartarus/Snowball/Ext/FrenchStemmer.cs | 10 +-
.../Tartarus/Snowball/Ext/ItalianStemmer.cs | 6 +-
.../Tartarus/Snowball/Ext/KpStemmer.cs | 4 +-
.../Tartarus/Snowball/Ext/PorterStemmer.cs | 4 +-
.../Tartarus/Snowball/Ext/PortugueseStemmer.cs | 6 +-
.../Tartarus/Snowball/Ext/RomanianStemmer.cs | 6 +-
.../Tartarus/Snowball/Ext/RussianStemmer.cs | 4 +-
.../Tartarus/Snowball/Ext/SpanishStemmer.cs | 6 +-
.../Tartarus/Snowball/Ext/TurkishStemmer.cs | 22 +-
.../Memory/DirectPostingsFormat.cs | 26 +-
src/Lucene.Net.Codecs/Memory/FSTOrdTermsReader.cs | 10 +-
src/Lucene.Net.Codecs/Memory/FSTTermsReader.cs | 8 +-
.../Pulsing/PulsingPostingsReader.cs | 2 +-
src/Lucene.Net.Expressions/JS/JavascriptLexer.cs | 40 +-
src/Lucene.Net.Expressions/JS/JavascriptParser.cs | 62 +--
src/Lucene.Net.Facet/DrillSidewaysScorer.cs | 12 +-
.../Taxonomy/WriterCache/CollisionMap.cs | 2 +-
.../Support/ToChildBlockJoinQuery.cs | 10 +-
src/Lucene.Net.Join/ToChildBlockJoinQuery.cs | 26 +-
src/Lucene.Net.QueryParser/Classic/QueryParser.cs | 36 +-
.../Classic/QueryParserTokenManager.cs | 456 ++++++++++-----------
.../Standard/Parser/StandardSyntaxParser.cs | 25 +-
.../Surround/Parser/QueryParser.cs | 17 +-
.../Surround/Parser/QueryParserTokenManager.cs | 22 +-
.../Prefix/AbstractVisitingPrefixTreeFilter.cs | 9 +-
src/Lucene.Net.Suggest/Spell/SpellChecker.cs | 19 +-
.../Suggest/Analyzing/FreeTextSuggester.cs | 34 +-
src/Lucene.Net/Codecs/BlockTreeTermsReader.cs | 6 +-
src/Lucene.Net/Codecs/Compressing/LZ4.cs | 8 +-
src/Lucene.Net/Search/FieldCacheRangeFilter.cs | 4 +-
src/Lucene.Net/Search/SloppyPhraseScorer.cs | 5 +-
src/Lucene.Net/Store/CompoundFileWriter.cs | 4 +-
src/Lucene.Net/Util/WAH8DocIdSet.cs | 4 +-
39 files changed, 459 insertions(+), 472 deletions(-)
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Compound/Hyphenation/HyphenationTree.cs b/src/Lucene.Net.Analysis.Common/Analysis/Compound/Hyphenation/HyphenationTree.cs
index 56522747b..82f094da3 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Compound/Hyphenation/HyphenationTree.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Compound/Hyphenation/HyphenationTree.cs
@@ -99,7 +99,7 @@ namespace Lucene.Net.Analysis.Compound.Hyphenation
byte v = m_vspace[k++];
while (v != 0)
{
- char c = (char)(v.TripleShift(4) - 1 + '0');;
+ char c = (char)(v.TripleShift(4) - 1 + '0');
buf.Append(c);
c = (char)(v & 0x0f);
if (c == 0)
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Dictionary.cs b/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Dictionary.cs
index 2f0ff3732..c5039b934 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Dictionary.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Hunspell/Dictionary.cs
@@ -1338,7 +1338,7 @@ namespace Lucene.Net.Analysis.Hunspell
{
if (rawFlags.Length == 0)
{
- return Arrays.Empty<char>(); ; // LUCENENET: Optimized char[] creation
+ return Arrays.Empty<char>(); // LUCENENET: Optimized char[] creation
}
StringBuilder builder = new StringBuilder();
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Sinks/TeeSinkTokenFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Sinks/TeeSinkTokenFilter.cs
index 6cac12f7c..03f694210 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Sinks/TeeSinkTokenFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Sinks/TeeSinkTokenFilter.cs
@@ -167,7 +167,7 @@ namespace Lucene.Net.Analysis.Sinks
AttributeSource.State finalState = CaptureState();
foreach (WeakReference<SinkTokenStream> @ref in sinks)
{
- SinkTokenStream sink; ;
+ SinkTokenStream sink;
if (@ref.TryGetTarget(out sink))
{
sink.SetFinalState(finalState);
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/BasqueStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/BasqueStemmer.cs
index 3ebe0ff40..d3a2fba7d 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/BasqueStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/BasqueStemmer.cs
@@ -579,7 +579,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab7:;
+ golab7: {/* LUCENENET: intentionally blank */}
} while (false);
lab3:
goto lab1;
@@ -641,7 +641,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark pV, line 35
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/CatalanStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/CatalanStemmer.cs
index b4c24c50d..bce4bc1fd 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/CatalanStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/CatalanStemmer.cs
@@ -1120,7 +1120,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab2;
}
} while (false);
- lab3:;
+ lab3: {/* LUCENENET: intentionally blank */}
} while (false);
lab2:
m_cursor = m_limit - v_3;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/EnglishStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/EnglishStemmer.cs
index 2a468be91..849722981 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/EnglishStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/EnglishStemmer.cs
@@ -323,7 +323,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto replab3;
end_of_outer_loop: { }
}
- replab3:;
+ replab3: {/* LUCENENET: intentionally blank */}
} while (false);
//lab2: - not referenced
m_cursor = v_3;
@@ -402,7 +402,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab5:;
+ golab5: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark p1, line 42
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/FrenchStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/FrenchStemmer.cs
index 12c2a82f1..c883fab5d 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/FrenchStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/FrenchStemmer.cs
@@ -482,7 +482,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab4:;
+ golab4: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark pV, line 67
@@ -775,7 +775,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
// <-, line 100
SliceFrom("iqU");
} while (false);
- lab1:;
+ lab1: {/* LUCENENET: intentionally blank */};
} while (false);
lab0:
break;
@@ -1095,7 +1095,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
// <-, line 142
SliceFrom("iqU");
} while (false);
- lab12:;
+ lab12: {/* LUCENENET: intentionally blank */}
} while (false);
lab11:
break;
@@ -1704,7 +1704,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
// <-, line 235
SliceFrom("c");
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab8:
goto lab3;
@@ -1717,7 +1717,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab2;
}
} while (false);
- lab3:;
+ lab3: {/* LUCENENET: intentionally blank */}
} while (false);
lab2:
m_cursor = m_limit - v_3;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/ItalianStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/ItalianStemmer.cs
index 449df0c5b..4f1c68819 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/ItalianStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/ItalianStemmer.cs
@@ -542,7 +542,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab7:;
+ golab7: {/* LUCENENET: intentionally blank */}
} while (false);
lab3:
goto lab1;
@@ -600,7 +600,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark pV, line 62
@@ -1304,7 +1304,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab3;
}
} while (false);
- lab4:;
+ lab4: {/* LUCENENET: intentionally blank */}
} while (false);
lab3:
m_cursor = m_limit - v_4;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/KpStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/KpStemmer.cs
index ccc6fe447..170c04da5 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/KpStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/KpStemmer.cs
@@ -2166,7 +2166,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto replab2;
end_of_outer_loop: { }
}
- replab2:;
+ replab2: {/* LUCENENET: intentionally blank */}
} while (false);
//lab1: // not referenced
m_cursor = v_2;
@@ -2439,7 +2439,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto replab19;
end_of_outer_loop_2: { }
}
- replab19:;
+ replab19: {/* LUCENENET: intentionally blank */}
} while (false);
lab18:
m_cursor = v_18;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PorterStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PorterStemmer.cs
index fe33f5429..b0cc8acb0 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PorterStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PorterStemmer.cs
@@ -780,7 +780,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto replab2;
end_of_outer_loop: { }
}
- replab2:;
+ replab2: {/* LUCENENET: intentionally blank */}
} while (false);
//lab1: // not referenced
m_cursor = v_2;
@@ -1035,7 +1035,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto replab24;
end_of_outer_loop_2: { }
}
- replab24:;
+ replab24: {/* LUCENENET: intentionally blank */}
} while (false);
lab23:
m_cursor = v_18;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PortugueseStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PortugueseStemmer.cs
index 017913090..f141f92b1 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PortugueseStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/PortugueseStemmer.cs
@@ -421,7 +421,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab7:;
+ golab7: {/* LUCENENET: intentionally blank */}
} while (false);
lab3:
goto lab1;
@@ -479,7 +479,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark pV, line 54
@@ -1227,7 +1227,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab2;
}
} while (false);
- lab3:;
+ lab3: {/* LUCENENET: intentionally blank */}
} while (false);
lab2:
m_cursor = m_limit - v_3;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RomanianStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RomanianStemmer.cs
index a65a5b289..671558282 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RomanianStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RomanianStemmer.cs
@@ -480,7 +480,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab7:;
+ golab7: {/* LUCENENET: intentionally blank */}
} while (false);
lab3:
goto lab1;
@@ -538,7 +538,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark pV, line 48
@@ -1144,7 +1144,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab4;
}
} while (false);
- lab5:;
+ lab5: {/* LUCENENET: intentionally blank */}
} while (false);
lab4:
m_cursor = m_limit - v_5;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RussianStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RussianStemmer.cs
index ba2c779af..69d54cc90 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RussianStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/RussianStemmer.cs
@@ -772,9 +772,9 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab1;
}
} while (false);
- lab5:;
+ lab5: {/* LUCENENET: intentionally blank */}
} while (false);
- lab2:;
+ lab2: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
m_cursor = m_limit - v_4;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/SpanishStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/SpanishStemmer.cs
index 9dddd1be5..2392e8faa 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/SpanishStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/SpanishStemmer.cs
@@ -370,7 +370,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
}
- golab7:;
+ golab7: {/* LUCENENET: intentionally blank */}
} while (false);
lab3:
goto lab1;
@@ -428,7 +428,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor++;
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab1:
// setmark pV, line 41
@@ -1279,7 +1279,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab2;
}
} while (false);
- lab3:;
+ lab3: {/* LUCENENET: intentionally blank */}
} while (false);
lab2:
m_cursor = m_limit - v_3;
diff --git a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/TurkishStemmer.cs b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/TurkishStemmer.cs
index ee50ee2d4..a94f62503 100644
--- a/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/TurkishStemmer.cs
+++ b/src/Lucene.Net.Analysis.Common/Tartarus/Snowball/Ext/TurkishStemmer.cs
@@ -585,7 +585,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
}
m_cursor--;
}
- golab24:;
+ golab24: {/* LUCENENET: intentionally blank */}
} while (false);
lab2:
m_cursor = m_limit - v_1;
@@ -1658,7 +1658,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab14;
}
} while (false);
- lab15:;
+ lab15: {/* LUCENENET: intentionally blank */}
} while (false);
lab14:
// unset continue_stemming_noun_suffixes, line 320
@@ -1875,7 +1875,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab28;
}
} while (false);
- lab28:;
+ lab28: {/* LUCENENET: intentionally blank */}
} while (false);
lab0:
// ], line 328
@@ -2005,9 +2005,9 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab6;
}
} while (false);
- lab6:;
+ lab6: {/* LUCENENET: intentionally blank */}
} while (false);
- lab3:;
+ lab3: {/* LUCENENET: intentionally blank */}
} while (false);
lab2:
goto lab0;
@@ -2129,7 +2129,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab8;
}
} while (false);
- lab9:;
+ lab9: {/* LUCENENET: intentionally blank */}
} while (false);
lab8:
goto lab0;
@@ -2216,7 +2216,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
return false;
}
} while (false);
- lab15:;
+ lab15: {/* LUCENENET: intentionally blank */}
} while (false);
lab0:
return true;
@@ -2419,7 +2419,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab4;
}
} while (false);
- lab5:;
+ lab5: {/* LUCENENET: intentionally blank */}
} while (false);
lab4:
goto lab0;
@@ -2738,7 +2738,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab25;
}
} while (false);
- lab26:;
+ lab26: {/* LUCENENET: intentionally blank */}
} while (false);
lab25:
goto lab0;
@@ -2885,7 +2885,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab34;
}
} while (false);
- lab35:;
+ lab35: {/* LUCENENET: intentionally blank */}
} while (false);
lab34:
goto lab0;
@@ -3099,7 +3099,7 @@ namespace Lucene.Net.Tartarus.Snowball.Ext
goto lab53;
}
} while (false);
- lab53:;
+ lab53: {/* LUCENENET: intentionally blank */}
} while (false);
lab0:
return true;
diff --git a/src/Lucene.Net.Codecs/Memory/DirectPostingsFormat.cs b/src/Lucene.Net.Codecs/Memory/DirectPostingsFormat.cs
index c807fc77a..f995efa4b 100644
--- a/src/Lucene.Net.Codecs/Memory/DirectPostingsFormat.cs
+++ b/src/Lucene.Net.Codecs/Memory/DirectPostingsFormat.cs
@@ -48,7 +48,7 @@ namespace Lucene.Net.Codecs.Memory
using TermState = Index.TermState;
using Transition = Util.Automaton.Transition;
- // TODO:
+ // TODO:
// - build depth-N prefix hash?
// - or: longer dense skip lists than just next byte?
@@ -56,23 +56,23 @@ namespace Lucene.Net.Codecs.Memory
/// Wraps <see cref="Lucene41.Lucene41PostingsFormat"/> format for on-disk
/// storage, but then at read time loads and stores all
/// terms & postings directly in RAM as byte[], int[].
- ///
+ ///
/// <para><b>WARNING</b>: This is
/// exceptionally RAM intensive: it makes no effort to
/// compress the postings data, storing terms as separate
- /// byte[] and postings as separate int[], but as a result it
+ /// byte[] and postings as separate int[], but as a result it
/// gives substantial increase in search performance.
- ///
+ ///
/// </para>
/// <para>This postings format supports <see cref="TermsEnum.Ord"/>
/// and <see cref="TermsEnum.SeekExact(long)"/>.
- ///
+ ///
/// </para>
/// <para>Because this holds all term bytes as a single
/// byte[], you cannot have more than 2.1GB worth of term
/// bytes in a single segment.
/// </para>
- /// @lucene.experimental
+ /// @lucene.experimental
/// </summary>
[PostingsFormatName("Direct")] // LUCENENET specific - using PostingsFormatName attribute to ensure the default name passed from subclasses is the same as this class name
public sealed class DirectPostingsFormat : PostingsFormat
@@ -85,7 +85,7 @@ namespace Lucene.Net.Codecs.Memory
// TODO: allow passing/wrapping arbitrary postings format?
- public DirectPostingsFormat()
+ public DirectPostingsFormat()
: this(DEFAULT_MIN_SKIP_COUNT, DEFAULT_LOW_FREQ_CUTOFF)
{
}
@@ -95,9 +95,9 @@ namespace Lucene.Net.Codecs.Memory
/// same prefix before we put a skip pointer down. Terms
/// with docFreq less than or equal <paramref name="lowFreqCutoff"/> will use a single int[]
/// to hold all docs, freqs, position and offsets; terms
- /// with higher docFreq will use separate arrays.
+ /// with higher docFreq will use separate arrays.
/// </summary>
- public DirectPostingsFormat(int minSkipCount, int lowFreqCutoff)
+ public DirectPostingsFormat(int minSkipCount, int lowFreqCutoff)
: base()
{
_minSkipCount = minSkipCount;
@@ -1318,7 +1318,7 @@ namespace Lucene.Net.Codecs.Memory
// System.out.println(" beyond end; no terms will match");
// }
return;
- nextLabelContinue: ;
+ nextLabelContinue: {/* LUCENENET: intentionally blank */}
}
//nextLabelBreak: ; // LUCENENET NOTE: Not used
}
@@ -1639,7 +1639,7 @@ namespace Lucene.Net.Codecs.Memory
// }
}
}
- nextTermContinue:;
+ nextTermContinue: {/* LUCENENET: intentionally blank */}
}
//nextTermBreak: ; // LUCENENET NOTE: Not used
@@ -2282,7 +2282,7 @@ namespace Lucene.Net.Codecs.Memory
public override int Advance(int target)
{
-
+
//System.out.println(" advance target=" + target + " cur=" + docID() + " upto=" + upto + " of " + docIDs.length);
// if (DEBUG) {
// System.out.println("advance target=" + target + " len=" + docIDs.length);
@@ -2594,7 +2594,7 @@ namespace Lucene.Net.Codecs.Memory
{
if (payloads is null)
return null;
-
+
var payloadBytes = payloads[upto][posUpto/(hasOffsets ? 3 : 1)];
if (payloadBytes is null)
{
diff --git a/src/Lucene.Net.Codecs/Memory/FSTOrdTermsReader.cs b/src/Lucene.Net.Codecs/Memory/FSTOrdTermsReader.cs
index 8390aa696..9e0e55e67 100644
--- a/src/Lucene.Net.Codecs/Memory/FSTOrdTermsReader.cs
+++ b/src/Lucene.Net.Codecs/Memory/FSTOrdTermsReader.cs
@@ -35,7 +35,7 @@ namespace Lucene.Net.Codecs.Memory
/// <summary>
/// FST-based terms dictionary reader.
/// <para/>
- /// The FST index maps each term and its ord, and during seek
+ /// The FST index maps each term and its ord, and during seek
/// the ord is used fetch metadata from a single block.
/// The term dictionary is fully memory resident.
/// <para/>
@@ -566,7 +566,7 @@ namespace Lucene.Net.Codecs.Memory
private bool pending;
/// <summary>
- /// stack to record how current term is constructed,
+ /// stack to record how current term is constructed,
/// used to accumulate metadata or rewind term:
/// level == term.length + 1,
/// == 0 when term is null
@@ -695,7 +695,7 @@ namespace Lucene.Net.Codecs.Memory
frame = PopFrame();
}
return false;
- DFSContinue:;
+ DFSContinue: {/* LUCENENET: intentionally blank */}
}
DFSBreak:
DecodeStats();
@@ -816,8 +816,8 @@ namespace Lucene.Net.Codecs.Memory
}
/// <summary>
- /// Load frame for target arc(node) on fst, so that
- /// arc.label >= label and !fsa.reject(arc.label)
+ /// Load frame for target arc(node) on fst, so that
+ /// arc.label >= label and !fsa.reject(arc.label)
/// </summary>
private Frame LoadCeilFrame(int label, Frame top, Frame frame)
{
diff --git a/src/Lucene.Net.Codecs/Memory/FSTTermsReader.cs b/src/Lucene.Net.Codecs/Memory/FSTTermsReader.cs
index 6bd5b555d..aa59d93da 100644
--- a/src/Lucene.Net.Codecs/Memory/FSTTermsReader.cs
+++ b/src/Lucene.Net.Codecs/Memory/FSTTermsReader.cs
@@ -52,7 +52,7 @@ namespace Lucene.Net.Codecs.Memory
/// <summary>
/// FST-based terms dictionary reader.
/// <para/>
- /// The FST directly maps each term and its metadata,
+ /// The FST directly maps each term and its metadata,
/// it is memory resident.
/// <para/>
/// @lucene.experimental
@@ -612,7 +612,7 @@ namespace Lucene.Net.Codecs.Memory
frame = PopFrame();
}
return false;
- DFSContinue:;
+ DFSContinue: {/* LUCENENET: intentionally blank */}
}
DFSBreak:
LoadMetaData();
@@ -730,8 +730,8 @@ namespace Lucene.Net.Codecs.Memory
}
/// <summary>
- /// Load frame for target arc(node) on fst, so that
- /// arc.label >= label and !fsa.reject(arc.label)
+ /// Load frame for target arc(node) on fst, so that
+ /// arc.label >= label and !fsa.reject(arc.label)
/// </summary>
private Frame LoadCeilFrame(int label, Frame top, Frame frame)
{
diff --git a/src/Lucene.Net.Codecs/Pulsing/PulsingPostingsReader.cs b/src/Lucene.Net.Codecs/Pulsing/PulsingPostingsReader.cs
index fc57f35a6..75363e18e 100644
--- a/src/Lucene.Net.Codecs/Pulsing/PulsingPostingsReader.cs
+++ b/src/Lucene.Net.Codecs/Pulsing/PulsingPostingsReader.cs
@@ -381,7 +381,7 @@ namespace Lucene.Net.Codecs.Pulsing
}
else
{
- _accum += code.TripleShift(1); ; // shift off low bit
+ _accum += code.TripleShift(1); // shift off low bit
_freq = (code & 1) != 0 ? 1 : _postings.ReadVInt32();
// LUCENENET specific - to avoid boxing, changed from CompareTo() to IndexOptionsComparer.Compare()
diff --git a/src/Lucene.Net.Expressions/JS/JavascriptLexer.cs b/src/Lucene.Net.Expressions/JS/JavascriptLexer.cs
index 8a635a9af..7bd62dc69 100644
--- a/src/Lucene.Net.Expressions/JS/JavascriptLexer.cs
+++ b/src/Lucene.Net.Expressions/JS/JavascriptLexer.cs
@@ -148,7 +148,7 @@ namespace Lucene.Net.Expressions.JS
public override string GrammarFileName => "";
// $ANTLR start "AT_ADD"
-
+
public void MAT_ADD()
{
try
@@ -798,9 +798,9 @@ namespace Lucene.Net.Expressions.JS
goto loop1_break;
}
}
-//loop1_continue: ; // LUCENENET NOTE: Not used
+//loop1_continue: // LUCENENET NOTE: Not used
}
-loop1_break: ;
+loop1_break: {/* LUCENENET: intentionally blank */}
}
state.type = _type;
state.channel = _channel;
@@ -866,9 +866,9 @@ loop1_break: ;
goto loop2_break;
}
}
-//loop2_continue: ; // LUCENENET NOTE: Not used
+//loop2_continue: // LUCENENET NOTE: Not used
}
-loop2_break: ;
+loop2_break: {/* LUCENENET: intentionally blank */}
}
}
finally
@@ -929,9 +929,9 @@ loop2_break: ;
}
}
cnt3++;
- //loop3_continue: ; // LUCENENET NOTE: Not used
+ //loop3_continue: // LUCENENET NOTE: Not used
}
- loop3_break: ;
+ loop3_break:
Skip();
}
state.type = _type;
@@ -993,9 +993,9 @@ loop2_break: ;
goto loop4_break;
}
}
- //loop4_continue: ; // LUCENENET NOTE: Not used
+ //loop4_continue: // LUCENENET NOTE: Not used
}
- loop4_break: ;
+ loop4_break:
// src/java/org/apache/lucene/expressions/js/Javascript.g:347:43: ( EXPONENT )?
int alt5 = 2;
int LA5_0 = input.LA(1);
@@ -1058,9 +1058,9 @@ loop2_break: ;
}
}
cnt6++;
- //loop6_continue: ; // LUCENENET NOTE: Not used
+ //loop6_continue: // LUCENENET NOTE: Not used
}
- loop6_break: ;
+ loop6_break:
// src/java/org/apache/lucene/expressions/js/Javascript.g:348:28: ( EXPONENT )?
int alt7 = 2;
int LA7_0 = input.LA(1);
@@ -1164,9 +1164,9 @@ loop2_break: ;
}
}
cnt10++;
- //loop10_continue: ; // LUCENENET NOTE: Not used
+ //loop10_continue: // LUCENENET NOTE: Not used
}
- loop10_break: ;
+ loop10_break: {/* LUCENENET: intentionally blank */}
}
state.type = _type;
state.channel = _channel;
@@ -1283,9 +1283,9 @@ loop2_break: ;
}
}
cnt12++;
- //loop12_continue: ; // LUCENENET NOTE: Not used
+ //loop12_continue: // LUCENENET NOTE: Not used
}
- loop12_break: ;
+ loop12_break: {/* LUCENENET: intentionally blank */}
}
state.type = _type;
state.channel = _channel;
@@ -1367,9 +1367,9 @@ loop2_break: ;
goto loop13_break;
}
}
- //loop13_continue: ; // LUCENENET NOTE: Not used
+ //loop13_continue: // LUCENENET NOTE: Not used
}
- loop13_break: ;
+ loop13_break:
break;
}
}
@@ -1464,9 +1464,9 @@ loop2_break: ;
}
}
cnt16++;
- //loop16_continue: ; // LUCENENET NOTE: Not used
+ //loop16_continue: // LUCENENET NOTE: Not used
}
- loop16_break: ;
+ loop16_break: {/* LUCENENET: intentionally blank */}
}
}
finally
@@ -1561,7 +1561,7 @@ loop2_break: ;
// do for sure before leaving
// $ANTLR end "OCTALDIGIT"
-
+
public override void mTokens()
{
// src/java/org/apache/lucene/expressions/js/Javascript.g:1:8: ( AT_ADD | AT_BIT_AND | AT_BIT_NOT | AT_BIT_OR | AT_BIT_SHL | AT_BIT_SHR | AT_BIT_SHU | AT_BIT_XOR | AT_BOOL_AND | AT_BOOL_NOT | AT_BOOL_OR | AT_COLON | AT_COMMA | AT_COMP_EQ | AT_COMP_GT | AT_COMP_GTE | AT_COMP_LT | AT_COMP_LTE | AT_COMP_NEQ | AT_COND_QUE | AT_DIVIDE | AT_DOT | AT_LPAREN | AT_MODULO | AT_MULTIPLY | AT_RPAREN | AT_SUBTRACT | NAMESPACE_ID | WS | DECIMAL | OCTAL | HEX )
diff --git a/src/Lucene.Net.Expressions/JS/JavascriptParser.cs b/src/Lucene.Net.Expressions/JS/JavascriptParser.cs
index 8f9424f31..528b6a88f 100644
--- a/src/Lucene.Net.Expressions/JS/JavascriptParser.cs
+++ b/src/Lucene.Net.Expressions/JS/JavascriptParser.cs
@@ -527,9 +527,9 @@ namespace Lucene.Net.Expressions.JS
goto loop2_break;
}
}
- //loop2_continue: ; // LUCENENET NOTE: Not used
+ //loop2_continue: // LUCENENET NOTE: Not used
}
- loop2_break: ;
+ loop2_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root_0);
@@ -597,9 +597,9 @@ namespace Lucene.Net.Expressions.JS
goto loop3_break;
}
}
- //loop3_continue: ; // LUCENENET NOTE: Not used
+ //loop3_continue: // LUCENENET NOTE: Not used
}
- loop3_break: ;
+ loop3_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -662,9 +662,9 @@ namespace Lucene.Net.Expressions.JS
goto loop4_break;
}
}
- //loop4_continue: ; // LUCENENET NOTE: Not used
+ //loop4_continue: // LUCENENET NOTE: Not used
}
- loop4_break: ;
+ loop4_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -727,9 +727,9 @@ namespace Lucene.Net.Expressions.JS
goto loop5_break;
}
}
- //loop5_continue: ; // LUCENENET NOTE: Not used
+ //loop5_continue: // LUCENENET NOTE: Not used
}
- loop5_break: ;
+ loop5_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -794,9 +794,9 @@ namespace Lucene.Net.Expressions.JS
goto loop6_break;
}
}
- //loop6_continue: ; // LUCENENET NOTE: Not used
+ //loop6_continue: // LUCENENET NOTE: Not used
}
- loop6_break: ;
+ loop6_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -870,9 +870,9 @@ namespace Lucene.Net.Expressions.JS
goto loop7_break;
}
}
- //loop7_continue: ; // LUCENENET NOTE: Not used
+ //loop7_continue: // LUCENENET NOTE: Not used
}
- loop7_break: ;
+ loop7_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -948,9 +948,9 @@ namespace Lucene.Net.Expressions.JS
goto loop8_break;
}
}
- //loop8_continue: ; // LUCENENET NOTE: Not used
+ //loop8_continue: // LUCENENET NOTE: Not used
}
- loop8_break: ;
+ loop8_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -1026,9 +1026,9 @@ namespace Lucene.Net.Expressions.JS
goto loop9_break;
}
}
- //loop9_continue: ; // LUCENENET NOTE: Not used
+ //loop9_continue: // LUCENENET NOTE: Not used
}
- loop9_break: ;
+ loop9_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -1104,9 +1104,9 @@ namespace Lucene.Net.Expressions.JS
goto loop10_break;
}
}
- //loop10_continue: ; // LUCENENET NOTE: Not used
+ //loop10_continue: // LUCENENET NOTE: Not used
}
- loop10_break: ;
+ loop10_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -1182,9 +1182,9 @@ namespace Lucene.Net.Expressions.JS
goto loop11_break;
}
}
- //loop11_continue: ; // LUCENENET NOTE: Not used
+ //loop11_continue: // LUCENENET NOTE: Not used
}
- loop11_break: ;
+ loop11_break: {/* LUCENENET: intentionally blank */}
}
retval.Stop = input.LT(-1);
retval.Tree = (CommonTree)m_adaptor.RulePostProcessing(root);
@@ -1351,12 +1351,12 @@ namespace Lucene.Net.Expressions.JS
);
stream_AT_SUBTRACT.Add(AT_SUBTRACT43);
// AST REWRITE
- // elements:
- // token labels:
+ // elements:
+ // token labels:
// rule labels: retval
- // token list labels:
- // rule list labels:
- // wildcard labels:
+ // token list labels:
+ // rule list labels:
+ // wildcard labels:
retval.Tree = root;
RewriteRuleSubtreeStream stream_retval = new RewriteRuleSubtreeStream(m_adaptor, "rule retval"
, retval != null ? ((CommonTree)retval.Tree) : null);
@@ -1491,11 +1491,11 @@ namespace Lucene.Net.Expressions.JS
streamArguments.Add(arguments48.Tree);
// AST REWRITE
// elements: NAMESPACE_ID, arguments
- // token labels:
+ // token labels:
// rule labels: retval
- // token list labels:
- // rule list labels:
- // wildcard labels:
+ // token list labels:
+ // rule list labels:
+ // wildcard labels:
retval.Tree = root;
RewriteRuleSubtreeStream stream_retval = new RewriteRuleSubtreeStream(m_adaptor, "rule retval"
, retval != null ? ((CommonTree)retval.Tree) : null);
@@ -1698,9 +1698,9 @@ namespace Lucene.Net.Expressions.JS
goto loop16_break;
}
}
- //loop16_continue: ; // LUCENENET NOTE: Not used
+ //loop16_continue: // LUCENENET NOTE: Not used
}
- loop16_break: ;
+ loop16_break:
break;
}
}
diff --git a/src/Lucene.Net.Facet/DrillSidewaysScorer.cs b/src/Lucene.Net.Facet/DrillSidewaysScorer.cs
index 38b0919ce..14f7d5e4b 100644
--- a/src/Lucene.Net.Facet/DrillSidewaysScorer.cs
+++ b/src/Lucene.Net.Facet/DrillSidewaysScorer.cs
@@ -54,7 +54,7 @@ namespace Lucene.Net.Facet
private int collectDocID = -1;
private float collectScore;
- internal DrillSidewaysScorer(AtomicReaderContext context, Scorer baseScorer,
+ internal DrillSidewaysScorer(AtomicReaderContext context, Scorer baseScorer,
ICollector drillDownCollector, DocsAndCost[] dims, bool scoreSubDocsAtOnce)
{
this.dims = dims;
@@ -186,9 +186,9 @@ namespace Lucene.Net.Facet
/// drilldowns, or when the docs must be scored at once
/// (i.e., like <see cref="Search.BooleanScorer2"/>, not <see cref="Search.BooleanScorer"/>). In
/// this case we just .Next() on base and .Advance() on
- /// the dim filters.
+ /// the dim filters.
/// </summary>
- private void DoQueryFirstScoring(ICollector collector, DocIdSetIterator[] disis,
+ private void DoQueryFirstScoring(ICollector collector, DocIdSetIterator[] disis,
ICollector[] sidewaysCollectors, IBits[] bits, ICollector[] bitsSidewaysCollectors)
{
//if (DEBUG) {
@@ -270,14 +270,14 @@ namespace Lucene.Net.Facet
}
docID = baseScorer.NextDoc();
- nextDocContinue:;
+ nextDocContinue: {/* LUCENENET: intentionally blank */}
}
- //nextDocBreak:; // Not referenced
+ //nextDocBreak: // Not referenced
}
/// <summary>
/// Used when drill downs are highly constraining vs
- /// baseQuery.
+ /// baseQuery.
/// </summary>
private void DoDrillDownAdvanceScoring(ICollector collector, DocIdSetIterator[] disis, ICollector[] sidewaysCollectors)
{
diff --git a/src/Lucene.Net.Facet/Taxonomy/WriterCache/CollisionMap.cs b/src/Lucene.Net.Facet/Taxonomy/WriterCache/CollisionMap.cs
index 458b3b68b..630abf80e 100644
--- a/src/Lucene.Net.Facet/Taxonomy/WriterCache/CollisionMap.cs
+++ b/src/Lucene.Net.Facet/Taxonomy/WriterCache/CollisionMap.cs
@@ -243,7 +243,7 @@ namespace Lucene.Net.Facet.Taxonomy.WriterCache
Entry e = this.next;
if (e is null)
{
- throw IllegalStateException.Create(this.GetType() + " cannot get next entry"); ;
+ throw IllegalStateException.Create(this.GetType() + " cannot get next entry");
}
Entry n = e.next;
diff --git a/src/Lucene.Net.Join/Support/ToChildBlockJoinQuery.cs b/src/Lucene.Net.Join/Support/ToChildBlockJoinQuery.cs
index bef8ba320..b19404a9c 100644
--- a/src/Lucene.Net.Join/Support/ToChildBlockJoinQuery.cs
+++ b/src/Lucene.Net.Join/Support/ToChildBlockJoinQuery.cs
@@ -31,7 +31,7 @@ namespace Lucene.Net.Join
/// query joins in reverse: you provide a <see cref="Query"/> matching
/// parent documents and it joins down to child
/// documents.
- ///
+ ///
/// @lucene.experimental
/// </summary>
[Obsolete("Use Lucene.Net.Search.Join.ToChildBlockJoinQuery instead. This class will be removed in 4.8.0 release candidate."), System.ComponentModel.EditorBrowsable(System.ComponentModel.EditorBrowsableState.Never)]
@@ -39,7 +39,7 @@ namespace Lucene.Net.Join
{
/// <summary>
/// Message thrown from <see cref="ToChildBlockJoinScorer.ValidateParentDoc"/>
- /// on mis-use, when the parent query incorrectly returns child docs.
+ /// on mis-use, when the parent query incorrectly returns child docs.
/// </summary>
internal const string INVALID_QUERY_MESSAGE = "Parent query yields document which is not matched by parents filter, docID=";
@@ -58,7 +58,7 @@ namespace Lucene.Net.Join
/// Create a <see cref="ToChildBlockJoinQuery"/>.
/// </summary>
/// <param name="parentQuery"><see cref="Query"/> that matches parent documents</param>
- /// <param name="parentsFilter"><see cref="Filter"/> (must produce FixedBitSet per-segment, like <see cref="FixedBitSetCachingWrapperFilter"/>)
+ /// <param name="parentsFilter"><see cref="Filter"/> (must produce FixedBitSet per-segment, like <see cref="FixedBitSetCachingWrapperFilter"/>)
/// identifying the parent documents.</param>
/// <param name="doScores">True if parent scores should be calculated.</param>
public ToChildBlockJoinQuery(Query parentQuery, Filter parentsFilter, bool doScores)
@@ -256,12 +256,12 @@ namespace Lucene.Net.Join
}
//System.out.println(" " + childDoc);
return _childDoc;
- nextChildDocContinue:;
+ nextChildDocContinue: {/* LUCENENET: intentionally blank */}
}
}
/// <summary>
- /// Detect mis-use, where provided parent query in fact sometimes returns child documents.
+ /// Detect mis-use, where provided parent query in fact sometimes returns child documents.
/// </summary>
private void ValidateParentDoc()
{
diff --git a/src/Lucene.Net.Join/ToChildBlockJoinQuery.cs b/src/Lucene.Net.Join/ToChildBlockJoinQuery.cs
index 21e3ac7f2..da65d2aac 100644
--- a/src/Lucene.Net.Join/ToChildBlockJoinQuery.cs
+++ b/src/Lucene.Net.Join/ToChildBlockJoinQuery.cs
@@ -30,14 +30,14 @@ namespace Lucene.Net.Search.Join
/// query joins in reverse: you provide a <see cref="Query"/> matching
/// parent documents and it joins down to child
/// documents.
- ///
+ ///
/// @lucene.experimental
/// </summary>
public class ToChildBlockJoinQuery : Query
{
/// <summary>
/// Message thrown from <see cref="ToChildBlockJoinScorer.ValidateParentDoc"/>
- /// on mis-use, when the parent query incorrectly returns child docs.
+ /// on mis-use, when the parent query incorrectly returns child docs.
/// </summary>
internal const string INVALID_QUERY_MESSAGE = "Parent query yields document which is not matched by parents filter, docID=";
@@ -56,7 +56,7 @@ namespace Lucene.Net.Search.Join
/// Create a <see cref="ToChildBlockJoinQuery"/>.
/// </summary>
/// <param name="parentQuery"><see cref="Query"/> that matches parent documents</param>
- /// <param name="parentsFilter"><see cref="Filter"/> (must produce FixedBitSet per-segment, like <see cref="FixedBitSetCachingWrapperFilter"/>)
+ /// <param name="parentsFilter"><see cref="Filter"/> (must produce FixedBitSet per-segment, like <see cref="FixedBitSetCachingWrapperFilter"/>)
/// identifying the parent documents.</param>
/// <param name="doScores">True if parent scores should be calculated.</param>
public ToChildBlockJoinQuery(Query parentQuery, Filter parentsFilter, bool doScores)
@@ -68,7 +68,7 @@ namespace Lucene.Net.Search.Join
_doScores = doScores;
}
- private ToChildBlockJoinQuery(Query origParentQuery, Query parentQuery, Filter parentsFilter, bool doScores)
+ private ToChildBlockJoinQuery(Query origParentQuery, Query parentQuery, Filter parentsFilter, bool doScores)
: base()
{
_origParentQuery = origParentQuery;
@@ -76,7 +76,7 @@ namespace Lucene.Net.Search.Join
_parentsFilter = parentsFilter;
_doScores = doScores;
}
-
+
public override Weight CreateWeight(IndexSearcher searcher)
{
return new ToChildBlockJoinWeight(this, _parentQuery.CreateWeight(searcher), _parentsFilter, _doScores);
@@ -89,7 +89,7 @@ namespace Lucene.Net.Search.Join
private readonly Filter _parentsFilter;
private readonly bool _doScores;
- public ToChildBlockJoinWeight(Query joinQuery, Weight parentWeight, Filter parentsFilter, bool doScores)
+ public ToChildBlockJoinWeight(Query joinQuery, Weight parentWeight, Filter parentsFilter, bool doScores)
: base()
{
_joinQuery = joinQuery;
@@ -140,7 +140,7 @@ namespace Lucene.Net.Search.Join
return new ToChildBlockJoinScorer(this, parentScorer, (FixedBitSet)parents, _doScores, acceptDocs);
}
-
+
public override Explanation Explain(AtomicReaderContext reader, int doc)
{
// TODO
@@ -163,7 +163,7 @@ namespace Lucene.Net.Search.Join
private int _childDoc = -1;
private int _parentDoc;
- public ToChildBlockJoinScorer(Weight weight, Scorer parentScorer, FixedBitSet parentBits, bool doScores, IBits acceptDocs)
+ public ToChildBlockJoinScorer(Weight weight, Scorer parentScorer, FixedBitSet parentBits, bool doScores, IBits acceptDocs)
: base(weight)
{
_doScores = doScores;
@@ -176,7 +176,7 @@ namespace Lucene.Net.Search.Join
{
return new JCG.List<ChildScorer> { new ChildScorer(_parentScorer, "BLOCK_JOIN") };
}
-
+
public override int NextDoc()
{
//System.out.println("Q.nextDoc() parentDoc=" + parentDoc + " childDoc=" + childDoc);
@@ -254,12 +254,12 @@ namespace Lucene.Net.Search.Join
}
//System.out.println(" " + childDoc);
return _childDoc;
- nextChildDocContinue:;
+ nextChildDocContinue: {/* LUCENENET: intentionally blank */}
}
}
/// <summary>
- /// Detect mis-use, where provided parent query in fact sometimes returns child documents.
+ /// Detect mis-use, where provided parent query in fact sometimes returns child documents.
/// </summary>
private void ValidateParentDoc()
{
@@ -275,7 +275,7 @@ namespace Lucene.Net.Search.Join
{
return _parentScore;
}
-
+
public override int Freq => _parentFreq;
public override int Advance(int childTarget)
@@ -334,7 +334,7 @@ namespace Lucene.Net.Search.Join
{
_parentQuery.ExtractTerms(terms);
}
-
+
public override Query Rewrite(IndexReader reader)
{
Query parentRewrite = _parentQuery.Rewrite(reader);
diff --git a/src/Lucene.Net.QueryParser/Classic/QueryParser.cs b/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
index cb8e859f6..8bfaa1179 100644
--- a/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
+++ b/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
@@ -40,7 +40,7 @@ namespace Lucene.Net.QueryParsers.Classic
/// <item><description> a term followed by a colon, indicating the field to be searched.
/// This enables one to construct queries which search multiple fields.</description></item>
/// </list>
- ///
+ ///
/// <para/>
/// A clause may be either:
/// <list type="bullet">
@@ -49,20 +49,20 @@ namespace Lucene.Net.QueryParsers.Classic
/// with a <c>+</c>/<c>-</c> prefix to require any of a set of
/// terms.</description></item>
/// </list>
- ///
+ ///
/// <para/>
/// Thus, in BNF, the query grammar is:
/// <code>
/// Query ::= ( Clause )*
/// Clause ::= ["+", "-"] [<TERM> ":"] ( <TERM> | "(" Query ")" )
/// </code>
- ///
+ ///
/// <para>
/// Examples of appropriately formatted queries can be found in the <a
/// href="../../../../../../queryparsersyntax.html">query syntax
/// documentation</a>.
/// </para>
- ///
+ ///
/// <para>
/// In <see cref="TermRangeQuery" />s, QueryParser tries to detect date values, e.g.
/// <tt>date:[6/1/2005 TO 6/4/2005]</tt> produces a range query that searches
@@ -85,14 +85,14 @@ namespace Lucene.Net.QueryParsers.Classic
/// <see cref="QueryParserBase.GetRangeQuery(string, string, string, bool, bool)" /> to
/// use a different method for date conversion.
/// </para>
- ///
+ ///
/// <para>Note that <see cref="QueryParser"/> is <em>not</em> thread-safe.</para>
- ///
+ ///
/// <para><b>NOTE</b>: there is a new QueryParser in contrib, which matches
/// the same syntax as this class, but is more modular,
/// enabling substantial customization to how a query is created.
/// </para>
- ///
+ ///
/// <b>NOTE</b>: You must specify the required <see cref="LuceneVersion" /> compatibility when
/// creating QueryParser:
/// <list type="bullet">
@@ -104,7 +104,7 @@ namespace Lucene.Net.QueryParsers.Classic
{
// NOTE: This was moved into the QueryParserBase class.
- // * The default operator for parsing queries.
+ // * The default operator for parsing queries.
// * Use <see cref="QueryParser.DefaultOperator"/> to change it.
// */
@@ -290,7 +290,7 @@ namespace Lucene.Net.QueryParsers.Classic
}
else
{
- ;
+ /* LUCENENET: intentionally blank */
}
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
@@ -622,19 +622,19 @@ namespace Lucene.Net.QueryParsers.Classic
//private static void Jj_la1_init_0()
//{
// jj_la1_0 = new uint[]
- // {
- // 0x300, 0x300, 0x1c00, 0x1c00, 0xfda7f00, 0x120000, 0x40000, 0xfda6000, 0x9d22000, 0x200000,
- // 0x200000, 0x40000, 0x6000000, 0x80000000, 0x10000000, 0x80000000, 0x60000000, 0x40000,
- // 0x200000, 0x40000, 0xfda2000,
+ // {
+ // 0x300, 0x300, 0x1c00, 0x1c00, 0xfda7f00, 0x120000, 0x40000, 0xfda6000, 0x9d22000, 0x200000,
+ // 0x200000, 0x40000, 0x6000000, 0x80000000, 0x10000000, 0x80000000, 0x60000000, 0x40000,
+ // 0x200000, 0x40000, 0xfda2000,
// };
//}
//private static void Jj_la1_init_1()
//{
// jj_la1_1 = new int[]
- // {
- // 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x1, 0x0,
- // 0x1, 0x0, 0x0, 0x0, 0x0, 0x0,
+ // {
+ // 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x1, 0x0,
+ // 0x1, 0x0, 0x0, 0x0, 0x0, 0x0,
// };
//}
@@ -715,7 +715,7 @@ namespace Lucene.Net.QueryParsers.Classic
throw GenerateParseException();
}
- // LUCENENET: It is no longer good practice to use binary serialization.
+ // LUCENENET: It is no longer good practice to use binary serialization.
// See: https://github.com/dotnet/corefx/issues/23584#issuecomment-325724568
#if FEATURE_SERIALIZABLE_EXCEPTIONS
[Serializable]
@@ -838,7 +838,7 @@ namespace Lucene.Net.QueryParsers.Classic
jj_expentries.Add(jj_expentry);
goto jj_entries_loop_break;
}
- jj_entries_loop_continue: ;
+ jj_entries_loop_continue: {/* LUCENENET: intentionally blank */}
}
jj_entries_loop_break:
if (pos != 0) jj_lasttokens[(jj_endpos = pos) - 1] = kind;
diff --git a/src/Lucene.Net.QueryParser/Classic/QueryParserTokenManager.cs b/src/Lucene.Net.QueryParser/Classic/QueryParserTokenManager.cs
index a1b87821b..069879684 100644
--- a/src/Lucene.Net.QueryParser/Classic/QueryParserTokenManager.cs
+++ b/src/Lucene.Net.QueryParser/Classic/QueryParserTokenManager.cs
@@ -49,10 +49,10 @@ namespace Lucene.Net.QueryParsers.Classic
{
switch (pos)
{
-
- default:
+
+ default:
return - 1;
-
+
}
}
@@ -70,37 +70,37 @@ namespace Lucene.Net.QueryParsers.Classic
{
switch (m_curChar)
{
-
- case (char) (40):
+
+ case (char) (40):
return JjStopAtPos(0, 14);
-
- case (char) (41):
+
+ case (char) (41):
return JjStopAtPos(0, 15);
-
- case (char) (42):
+
+ case (char) (42):
return JjStartNfaWithStates_2(0, 17, 49);
-
+
case (char) (43):
return JjStartNfaWithStates_2(0, 11, 15);
-
+
case (char) (45):
return JjStartNfaWithStates_2(0, 12, 15);
-
- case (char) (58):
+
+ case (char) (58):
return JjStopAtPos(0, 16);
-
- case (char) (91):
+
+ case (char) (91):
return JjStopAtPos(0, 25);
-
- case (char) (94):
+
+ case (char) (94):
return JjStopAtPos(0, 18);
-
- case (char) (123):
+
+ case (char) (123):
return JjStopAtPos(0, 26);
-
+
default:
return JjMoveNfa_2(0, 0);
-
+
}
}
private int JjStartNfaWithStates_2(int pos, int kind, int state)
@@ -135,12 +135,12 @@ namespace Lucene.Net.QueryParsers.Classic
if (m_curChar < 64)
{
ulong l = (ulong) (1L << (int) m_curChar);
- do
+ do
{
switch (jjstateSet[--i])
{
-
- case 49:
+
+ case 49:
case 33:
if ((0xfbff7cf8ffffd9ffL & l) == (ulong)0L)
break;
@@ -148,7 +148,7 @@ namespace Lucene.Net.QueryParsers.Classic
kind = 23;
JjCheckNAddTwoStates(33, 34);
break;
-
+
case 0:
if ((0xfbff54f8ffffd9ffL & l) != (ulong)0L)
{
@@ -186,22 +186,22 @@ namespace Lucene.Net.QueryParsers.Classic
if (m_curChar == 38)
jjstateSet[jjnewStateCnt++] = 4;
break;
-
- case 4:
+
+ case 4:
if (m_curChar == 38 && kind > 8)
kind = 8;
break;
-
- case 5:
+
+ case 5:
if (m_curChar == 38)
jjstateSet[jjnewStateCnt++] = 4;
break;
-
- case 13:
+
+ case 13:
if (m_curChar == 33 && kind > 10)
kind = 10;
break;
-
+
case 14:
if ((0x280200000000L & l) != 0L)
jjstateSet[jjnewStateCnt++] = 15;
@@ -214,41 +214,41 @@ namespace Lucene.Net.QueryParsers.Classic
if (m_curChar == 34)
JjCheckNAddStates(3, 5);
break;
- case 17:
+ case 17:
if ((0xfffffffbffffffffL & l) != (ulong) 0L)
JjCheckNAddStates(3, 5);
break;
-
- case 19:
+
+ case 19:
JjCheckNAddStates(3, 5);
break;
-
- case 20:
+
+ case 20:
if (m_curChar == 34 && kind > 19)
kind = 19;
break;
-
- case 22:
+
+ case 22:
if ((0x3ff000000000000L & l) == 0L)
break;
if (kind > 21)
kind = 21;
JjCheckNAddStates(11, 14);
break;
-
- case 23:
+
+ case 23:
if (m_curChar == 46)
JjCheckNAdd(24);
break;
-
- case 24:
+
+ case 24:
if ((0x3ff000000000000L & l) == 0L)
break;
if (kind > 21)
kind = 21;
JjCheckNAddStates(15, 17);
break;
-
+
case 25:
if ((0x7bff78f8ffffd9ffL & l) == (ulong)0L)
break;
@@ -256,13 +256,13 @@ namespace Lucene.Net.QueryParsers.Classic
kind = 21;
JjCheckNAddTwoStates(25, 26);
break;
-
- case 27:
+
+ case 27:
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(25, 26);
break;
-
+
case 28:
if ((0x7bff78f8ffffd9ffL & l) == 0L)
break;
@@ -270,18 +270,18 @@ namespace Lucene.Net.QueryParsers.Classic
kind = 21;
JjCheckNAddTwoStates(28, 29);
break;
-
- case 30:
+
+ case 30:
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(28, 29);
break;
-
+
case 31:
if (m_curChar == 42 && kind > 22)
kind = 22;
break;
-
+
case 32:
if ((0xfbff54f8ffffd9ffL & l) == (ulong)0L)
break;
@@ -333,9 +333,9 @@ namespace Lucene.Net.QueryParsers.Classic
case 47:
JjCheckNAddStates(18, 20);
break;
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -346,13 +346,13 @@ namespace Lucene.Net.QueryParsers.Classic
// But changing it back made more tests pass, so I am working under the assumption 63
// is the correct value.
//ulong l = (ulong)(1L << (curChar & 077));
- ulong l = (ulong) (1L << (m_curChar & 63));
- do
+ ulong l = (ulong) (1L << (m_curChar & 63));
+ do
{
switch (jjstateSet[--i])
{
-
- case 49:
+
+ case 49:
if ((0x97ffffff87ffffffL & l) != (ulong) 0L)
{
if (kind > 23)
@@ -362,8 +362,8 @@ namespace Lucene.Net.QueryParsers.Classic
else if (m_curChar == 92)
JjCheckNAddTwoStates(35, 35);
break;
-
- case 0:
+
+ case 0:
if ((0x97ffffff87ffffffL & l) != (ulong) 0L)
{
if (kind > 20)
@@ -393,106 +393,106 @@ namespace Lucene.Net.QueryParsers.Classic
else if (m_curChar == 65)
jjstateSet[jjnewStateCnt++] = 2;
break;
-
- case 1:
+
+ case 1:
if (m_curChar == 68 && kind > 8)
kind = 8;
break;
-
- case 2:
+
+ case 2:
if (m_curChar == 78)
jjstateSet[jjnewStateCnt++] = 1;
break;
-
- case 3:
+
+ case 3:
if (m_curChar == 65)
jjstateSet[jjnewStateCnt++] = 2;
break;
-
- case 6:
+
+ case 6:
if (m_curChar == 82 && kind > 9)
kind = 9;
break;
-
- case 7:
+
+ case 7:
if (m_curChar == 79)
jjstateSet[jjnewStateCnt++] = 6;
break;
-
- case 8:
+
+ case 8:
if (m_curChar == 124 && kind > 9)
kind = 9;
break;
-
- case 9:
+
+ case 9:
if (m_curChar == 124)
jjstateSet[jjnewStateCnt++] = 8;
break;
-
- case 10:
+
+ case 10:
if (m_curChar == 84 && kind > 10)
kind = 10;
break;
-
- case 11:
+
+ case 11:
if (m_curChar == 79)
jjstateSet[jjnewStateCnt++] = 10;
break;
-
- case 12:
+
+ case 12:
if (m_curChar == 78)
jjstateSet[jjnewStateCnt++] = 11;
break;
-
- case 17:
+
+ case 17:
if ((0xffffffffefffffffL & l) != (ulong) 0L)
JjCheckNAddStates(3, 5);
break;
-
- case 18:
+
+ case 18:
if (m_curChar == 92)
jjstateSet[jjnewStateCnt++] = 19;
break;
-
- case 19:
+
+ case 19:
JjCheckNAddStates(3, 5);
break;
-
- case 21:
+
+ case 21:
if (m_curChar != 126)
break;
if (kind > 21)
kind = 21;
JjCheckNAddStates(24, 26);
break;
-
- case 25:
+
+ case 25:
if ((0x97ffffff87ffffffL & l) == (ulong) 0L)
break;
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(25, 26);
break;
-
- case 26:
+
+ case 26:
if (m_curChar == 92)
JjAddStates(27, 28);
break;
-
- case 27:
+
+ case 27:
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(25, 26);
break;
-
- case 28:
+
+ case 28:
if ((0x97ffffff87ffffffL & l) == (ulong) 0L)
break;
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(28, 29);
break;
-
+
case 29:
if (m_curChar == 92)
JjAddStates(29, 30);
@@ -572,7 +572,7 @@ namespace Lucene.Net.QueryParsers.Classic
break;
default: break;
-
+
}
}
while (i != startsAt);
@@ -584,21 +584,21 @@ namespace Lucene.Net.QueryParsers.Classic
ulong l1 = (ulong) (1L << (hiByte & 63));
int i2 = (m_curChar & 0xff) >> 6;
ulong l2 = (ulong) (1L << (m_curChar & 63));
- do
+ do
{
switch (jjstateSet[--i])
{
-
- case 49:
- case 33:
+
+ case 49:
+ case 33:
if (!JjCanMove_2(hiByte, i1, i2, l1, l2))
break;
if (kind > 23)
kind = 23;
JjCheckNAddTwoStates(33, 34);
break;
-
- case 0:
+
+ case 0:
if (JjCanMove_0(hiByte, i1, i2, l1, l2))
{
if (kind > 7)
@@ -617,34 +617,34 @@ namespace Lucene.Net.QueryParsers.Classic
JjCheckNAddStates(6, 10);
}
break;
-
- case 15:
+
+ case 15:
if (JjCanMove_0(hiByte, i1, i2, l1, l2) && kind > 13)
kind = 13;
break;
- case 17:
+ case 17:
case 19:
if (JjCanMove_1(hiByte, i1, i2, l1, l2))
JjCheckNAddStates(3, 5);
break;
-
- case 25:
+
+ case 25:
if (!JjCanMove_2(hiByte, i1, i2, l1, l2))
break;
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(25, 26);
break;
-
- case 27:
+
+ case 27:
if (!JjCanMove_1(hiByte, i1, i2, l1, l2))
break;
if (kind > 21)
kind = 21;
JjCheckNAddTwoStates(25, 26);
break;
-
- case 28:
+
+ case 28:
if (!JjCanMove_2(hiByte, i1, i2, l1, l2))
break;
if (kind > 21)
@@ -658,15 +658,15 @@ namespace Lucene.Net.QueryParsers.Classic
kind = 21;
JjCheckNAddTwoStates(28, 29);
break;
- case 32:
+ case 32:
if (!JjCanMove_2(hiByte, i1, i2, l1, l2))
break;
if (kind > 23)
kind = 23;
JjCheckNAddTwoStates(33, 34);
break;
-
- case 35:
+
+ case 35:
if (!JjCanMove_1(hiByte, i1, i2, l1, l2))
break;
if (kind > 23)
@@ -703,14 +703,14 @@ namespace Lucene.Net.QueryParsers.Classic
if (JjCanMove_2(hiByte, i1, i2, l1, l2))
JjCheckNAddStates(18, 20);
break;
-
- case 47:
+
+ case 47:
if (JjCanMove_1(hiByte, i1, i2, l1, l2))
JjCheckNAddStates(18, 20);
break;
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -752,34 +752,34 @@ namespace Lucene.Net.QueryParsers.Classic
if (m_curChar < 64)
{
ulong l = (ulong) (1L << (int) m_curChar);
- do
+ do
{
switch (jjstateSet[--i])
{
-
- case 0:
+
+ case 0:
if ((0x3ff000000000000L & l) == 0L)
break;
if (kind > 27)
kind = 27;
JjAddStates(31, 32);
break;
-
- case 1:
+
+ case 1:
if (m_curChar == 46)
JjCheckNAdd(2);
break;
-
- case 2:
+
+ case 2:
if ((0x3ff000000000000L & l) == 0L)
break;
if (kind > 27)
kind = 27;
JjCheckNAdd(2);
break;
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -791,9 +791,9 @@ namespace Lucene.Net.QueryParsers.Classic
{
switch (jjstateSet[--i])
{
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -805,13 +805,13 @@ namespace Lucene.Net.QueryParsers.Classic
long l1 = 1L << (hiByte & 63);
int i2 = (m_curChar & 0xff) >> 6;
long l2 = 1L << (m_curChar & 63);
- do
+ do
{
switch (jjstateSet[--i])
{
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -839,7 +839,7 @@ namespace Lucene.Net.QueryParsers.Classic
{
switch (pos)
{
-
+
case 0:
if ((active0 & 0x10000000L) != 0L)
{
@@ -847,10 +847,10 @@ namespace Lucene.Net.QueryParsers.Classic
return 6;
}
return - 1;
-
- default:
+
+ default:
return - 1;
-
+
}
}
private int JjStartNfa_1(int pos, long active0)
@@ -861,19 +861,19 @@ namespace Lucene.Net.QueryParsers.Classic
{
switch (m_curChar)
{
-
+
case (char)84:
return JjMoveStringLiteralDfa1_1(0x10000000L);
-
- case (char)93:
+
+ case (char)93:
return JjStopAtPos(0, 29);
case (char)125:
return JjStopAtPos(0, 30);
- default:
+ default:
return JjMoveNfa_1(0, 0);
-
+
}
}
private int JjMoveStringLiteralDfa1_1(long active0)
@@ -889,15 +889,15 @@ namespace Lucene.Net.QueryParsers.Classic
}
switch (m_curChar)
{
-
+
case (char) (79):
if ((active0 & 0x10000000L) != 0L)
return JjStartNfaWithStates_1(1, 28, 6);
break;
-
- default:
+
+ default:
break;
-
+
}
return JjStartNfa_1(0, active0);
}
@@ -929,11 +929,11 @@ namespace Lucene.Net.QueryParsers.Classic
if (m_curChar < 64)
{
ulong l = (ulong) (1L << (int) m_curChar);
- do
+ do
{
switch (jjstateSet[--i])
{
-
+
case 0:
if ((0xfffffffeffffffffL & l) != (ulong)0L)
{
@@ -949,27 +949,27 @@ namespace Lucene.Net.QueryParsers.Classic
else if (m_curChar == 34)
JjCheckNAddTwoStates(2, 4);
break;
-
- case 1:
+
+ case 1:
if (m_curChar == 34)
JjCheckNAddTwoStates(2, 4);
break;
-
+
case 2:
if ((0xfffffffbffffffffL & l) != (ulong)0L)
JjCheckNAddStates(33, 35);
break;
-
- case 3:
+
+ case 3:
if (m_curChar == 34)
JjCheckNAddStates(33, 35);
break;
-
- case 5:
+
+ case 5:
if (m_curChar == 34 && kind > 31)
kind = 31;
break;
-
+
case 6:
if ((0xfffffffeffffffffL & l) == (ulong)0L)
break;
@@ -977,9 +977,9 @@ namespace Lucene.Net.QueryParsers.Classic
kind = 32;
JjCheckNAdd(6);
break;
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -987,12 +987,12 @@ namespace Lucene.Net.QueryParsers.Classic
else if (m_curChar < 128)
{
ulong l = (ulong) (1L << (m_curChar & 63));
- do
+ do
{
switch (jjstateSet[--i])
{
-
- case 0:
+
+ case 0:
case 6:
if ((0xdfffffffdfffffffL & l) == (ulong)0L)
break;
@@ -1000,18 +1000,18 @@ namespace Lucene.Net.QueryParsers.Classic
kind = 32;
JjCheckNAdd(6);
break;
-
- case 2:
+
+ case 2:
JjAddStates(33, 35);
break;
-
- case 4:
+
+ case 4:
if (m_curChar == 92)
jjstateSet[jjnewStateCnt++] = 3;
break;
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -1023,12 +1023,12 @@ namespace Lucene.Net.QueryParsers.Classic
ulong l1 = (ulong) (1L << (hiByte & 63));
int i2 = (m_curChar & 0xff) >> 6;
ulong l2 = (ulong) (1L << (m_curChar & 63));
- do
+ do
{
switch (jjstateSet[--i])
{
-
- case 0:
+
+ case 0:
if (JjCanMove_0(hiByte, i1, i2, l1, l2))
{
if (kind > 7)
@@ -1041,22 +1041,22 @@ namespace Lucene.Net.QueryParsers.Classic
JjCheckNAdd(6);
}
break;
-
- case 2:
+
+ case 2:
if (JjCanMove_1(hiByte, i1, i2, l1, l2))
JjAddStates(33, 35);
break;
-
- case 6:
+
+ case 6:
if (!JjCanMove_1(hiByte, i1, i2, l1, l2))
break;
if (kind > 32)
kind = 32;
JjCheckNAdd(6);
break;
-
+
default: break;
-
+
}
}
while (i != startsAt);
@@ -1081,80 +1081,80 @@ namespace Lucene.Net.QueryParsers.Classic
}
}
internal static readonly int[] jjnextStates = new int[]{
- 37, 39, 40, 17, 18, 20, 42, 45, 31, 46, 43, 22, 23, 25, 26, 24,
- 25, 26, 45, 31, 46, 44, 47, 35, 22, 28, 29, 27, 27, 30, 30, 0,
+ 37, 39, 40, 17, 18, 20, 42, 45, 31, 46, 43, 22, 23, 25, 26, 24,
+ 25, 26, 45, 31, 46, 44, 47, 35, 22, 28, 29, 27, 27, 30, 30, 0,
1, 2, 4, 5
};
private static bool JjCanMove_0(int hiByte, int i1, int i2, ulong l1, ulong l2)
{
switch (hiByte)
{
-
- case 48:
+
+ case 48:
return ((jjbitVec0[i2] & l2) != (ulong) 0L);
-
- default:
+
+ default:
return false;
-
+
}
}
private static bool JjCanMove_1(int hiByte, int i1, int i2, ulong l1, ulong l2)
{
switch (hiByte)
{
-
- case 0:
+
+ case 0:
return ((jjbitVec3[i2] & l2) != (ulong) 0L);
-
- default:
+
+ default:
if ((jjbitVec1[i1] & l1) != (ulong) 0L)
return true;
return false;
-
+
}
}
private static bool JjCanMove_2(int hiByte, int i1, int i2, ulong l1, ulong l2)
{
switch (hiByte)
{
-
- case 0:
+
+ case 0:
return ((jjbitVec3[i2] & l2) != (ulong) 0L);
-
- case 48:
+
+ case 48:
return ((jjbitVec1[i2] & l2) != (ulong) 0L);
-
- default:
+
+ default:
if ((jjbitVec4[i1] & l1) != (ulong) 0L)
return true;
return false;
-
+
}
}
///// <summary>Token literal values. </summary>
- //public static readonly string[] jjstrLiteralImages = new string[] {
- // "", null, null, null, null, null, null, null, null, null, null, "\x002B", "\x002D",
- // "\x0028", "\x0029", "\x003A", "\x002A", "\x005E", null, null, null, null, null, "\x005B", "\x007B",
+ //public static readonly string[] jjstrLiteralImages = new string[] {
+ // "", null, null, null, null, null, null, null, null, null, null, "\x002B", "\x002D",
+ // "\x0028", "\x0029", "\x003A", "\x002A", "\x005E", null, null, null, null, null, "\x005B", "\x007B",
// null, "\x0054\x004F", "\x005D", null, null, "\x0054\x004F", "\x007D", null, null };
-
+
/// <summary>Token literal values. </summary>
public static readonly string[] jjstrLiteralImages = new string[]{
- "", null, null, null, null, null, null, null, null, null, null, "\x002B", "\x002D",
- null, "\x0028", "\x0029", "\x003A", "\x002A", "\x005E", null, null, null, null, null, null,
+ "", null, null, null, null, null, null, null, null, null, null, "\x002B", "\x002D",
+ null, "\x0028", "\x0029", "\x003A", "\x002A", "\x005E", null, null, null, null, null, null,
"\x005B", "\x007B", null, "\x0054\x004F", "\x005D", "\x007D", null, null };
-
+
/// <summary>Lexer state names. </summary>
public static readonly string[] lexStateNames = new string[] {
- "Boost",
- "Range",
+ "Boost",
+ "Range",
"DEFAULT"
};
-
+
/// <summary>Lex State array. </summary>
public static readonly int[] jjnewLexState = new int[] {
- -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, 0, -1, -1, -1, -1, -1, -1,
+ -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, 0, -1, -1, -1, -1, -1, -1,
1, 1, 2, -1, 2, 2, -1, -1
};
internal static readonly ulong[] jjtoToken = new ulong[] { 0x1ffffff01L };
@@ -1169,13 +1169,13 @@ namespace Lucene.Net.QueryParsers.Classic
InitBlock();
m_input_stream = stream;
}
-
+
/// <summary>Constructor. </summary>
public QueryParserTokenManager(ICharStream stream, int lexState):this(stream)
{
SwitchTo(lexState);
}
-
+
/// <summary>Reinitialise parser. </summary>
public virtual void ReInit(ICharStream stream)
{
@@ -1191,14 +1191,14 @@ namespace Lucene.Net.QueryParsers.Classic
for (i = 49; i-- > 0; )
jjrounds[i] = 0x80000000;
}
-
+
/// <summary>Reinitialise parser. </summary>
public virtual void ReInit(ICharStream stream, int lexState)
{
ReInit(stream);
SwitchTo(lexState);
}
-
+
/// <summary>Switch to specified lex state. </summary>
public virtual void SwitchTo(int lexState)
{
@@ -1207,7 +1207,7 @@ namespace Lucene.Net.QueryParsers.Classic
else
curLexState = lexState;
}
-
+
protected internal virtual Token JjFillToken()
{
Token t;
@@ -1223,29 +1223,29 @@ namespace Lucene.Net.QueryParsers.Classic
endLine = m_input_stream.EndLine;
endColumn = m_input_stream.EndColumn;
t = Token.NewToken(jjmatchedKind, curTokenImage);
-
+
t.BeginLine = beginLine;
t.EndLine = endLine;
t.BeginColumn = beginColumn;
t.EndColumn = endColumn;
-
+
return t;
}
-
+
internal int curLexState = 2;
internal int defaultLexState = 2;
internal int jjnewStateCnt;
internal uint jjround;
internal int jjmatchedPos;
internal int jjmatchedKind;
-
+
/// <summary>Get the next Token. </summary>
[SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
public virtual Token GetNextToken()
{
Token matchedToken;
int curPos = 0;
-
+
for (; ; )
{
try
@@ -1258,23 +1258,23 @@ namespace Lucene.Net.QueryParsers.Classic
matchedToken = JjFillToken();
return matchedToken;
}
-
+
switch (curLexState)
{
-
- case 0:
+
+ case 0:
jjmatchedKind = 0x7fffffff;
jjmatchedPos = 0;
curPos = JjMoveStringLiteralDfa0_0();
break;
-
- case 1:
+
+ case 1:
jjmatchedKind = 0x7fffffff;
jjmatchedPos = 0;
curPos = JjMoveStringLiteralDfa0_1();
break;
-
- case 2:
+
+ case 2:
jjmatchedKind = 0x7fffffff;
jjmatchedPos = 0;
curPos = JjMoveStringLiteralDfa0_2();
@@ -1325,10 +1325,10 @@ namespace Lucene.Net.QueryParsers.Classic
}
throw new TokenMgrError(EOFSeen, curLexState, error_line, error_column, error_after, m_curChar, TokenMgrError.LEXICAL_ERROR);
-EOFLoop: ;
+EOFLoop: {/* LUCENENET: intentionally blank */}
}
}
-
+
private void JjCheckNAdd(int state)
{
if (jjrounds[state] != jjround)
@@ -1339,7 +1339,7 @@ EOFLoop: ;
}
private void JjAddStates(int start, int end)
{
- do
+ do
{
jjstateSet[jjnewStateCnt++] = jjnextStates[start];
}
@@ -1350,10 +1350,10 @@ EOFLoop: ;
JjCheckNAdd(state1);
JjCheckNAdd(state2);
}
-
+
private void JjCheckNAddStates(int start, int end)
{
- do
+ do
{
JjCheckNAdd(jjnextStates[start]);
}
diff --git a/src/Lucene.Net.QueryParser/Flexible/Standard/Parser/StandardSyntaxParser.cs b/src/Lucene.Net.QueryParser/Flexible/Standard/Parser/StandardSyntaxParser.cs
index 215f6d3f7..866caf814 100644
--- a/src/Lucene.Net.QueryParser/Flexible/Standard/Parser/StandardSyntaxParser.cs
+++ b/src/Lucene.Net.QueryParser/Flexible/Standard/Parser/StandardSyntaxParser.cs
@@ -162,21 +162,21 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
// These changes were made to introduce operator precedence:
- // - Clause() now returns a QueryNode.
+ // - Clause() now returns a QueryNode.
// - The modifiers are consumed by Clause() and returned as part of the QueryNode Object
- // - Query does not consume conjunctions (AND, OR) anymore.
+ // - Query does not consume conjunctions (AND, OR) anymore.
// - This is now done by two new non-terminals: ConjClause and DisjClause
// The parse tree looks similar to this:
// Query ::= DisjQuery ( DisjQuery )*
- // DisjQuery ::= ConjQuery ( OR ConjQuery )*
+ // DisjQuery ::= ConjQuery ( OR ConjQuery )*
// ConjQuery ::= Clause ( AND Clause )*
- // Clause ::= [ Modifier ] ...
+ // Clause ::= [ Modifier ] ...
public IQueryNode Query(string field)
{
IList<IQueryNode> clauses = null;
IQueryNode c, first = null;
first = DisjQuery(field);
-
+
while (true)
{
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
@@ -191,7 +191,6 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
case RegexpToken.RANGEIN_START:
case RegexpToken.RANGEEX_START:
case RegexpToken.NUMBER:
- ;
break;
default:
jj_la1[4] = jj_gen;
@@ -222,13 +221,12 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
IQueryNode first, c;
IList<IQueryNode> clauses = null;
first = ConjQuery(field);
-
+
while (true)
{
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.OR:
- ;
break;
default:
jj_la1[5] = jj_gen;
@@ -260,13 +258,12 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
IQueryNode first, c;
IList<IQueryNode> clauses = null;
first = ModClause(field);
-
+
while (true)
{
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.AND:
- ;
break;
default:
jj_la1[6] = jj_gen;
@@ -305,7 +302,7 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
// mods=Modifiers() q=Clause(field)
// {
// if (mods == ModifierQueryNode.Modifier.MOD_NONE) firstQuery=q;
- //
+ //
// // do not create modifier nodes with MOD_NONE
// if (mods != ModifierQueryNode.Modifier.MOD_NONE) {
// q = new ModifierQueryNode(q, mods);
@@ -314,7 +311,7 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
// }
// (
// conj=Conjunction() mods=Modifiers() q=Clause(field)
- // {
+ // {
// // do not create modifier nodes with MOD_NONE
// if (mods != ModifierQueryNode.Modifier.MOD_NONE) {
// q = new ModifierQueryNode(q, mods);
@@ -501,7 +498,7 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
}
else
{
- ;
+ /* LUCENENET: intentionally blank */
}
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
@@ -1081,7 +1078,7 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard.Parser
throw GenerateParseException();
}
- // LUCENENET: It is no longer good practice to use binary serialization.
+ // LUCENENET: It is no longer good practice to use binary serialization.
// See: https://github.com/dotnet/corefx/issues/23584#issuecomment-325724568
#if FEATURE_SERIALIZABLE_EXCEPTIONS
[Serializable]
diff --git a/src/Lucene.Net.QueryParser/Surround/Parser/QueryParser.cs b/src/Lucene.Net.QueryParser/Surround/Parser/QueryParser.cs
index ebb35c25a..cbb86037c 100644
--- a/src/Lucene.Net.QueryParser/Surround/Parser/QueryParser.cs
+++ b/src/Lucene.Net.QueryParser/Surround/Parser/QueryParser.cs
@@ -53,7 +53,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
/// default, meaning the operands are adjacent, or may be provided as a
/// prefix from 2-99. So, for example, 3W(a, b) means that terms a and b
/// must appear within three positions of each other, or in other words, up
- /// to two terms may appear between a and b.
+ /// to two terms may appear between a and b.
/// </para>
/// </summary>
[SuppressMessage("Performance", "CA1822:Mark members as static", Justification = "This class is based on generated code")]
@@ -218,7 +218,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
{
if (Jj_2_1(2))
{
- ;
+ /* LUCENENET: intentionally blank */
}
else
{
@@ -250,7 +250,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.OR:
- ;
break;
default:
jj_la1[0] = jj_gen;
@@ -283,7 +282,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.AND:
- ;
break;
default:
jj_la1[1] = jj_gen;
@@ -316,7 +314,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.NOT:
- ;
break;
default:
jj_la1[2] = jj_gen;
@@ -349,7 +346,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.N:
- ;
break;
default:
jj_la1[3] = jj_gen;
@@ -380,7 +376,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.W:
- ;
break;
default:
jj_la1[4] = jj_gen;
@@ -487,7 +482,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.COMMA:
- ;
break;
default:
jj_la1[7] = jj_gen;
@@ -556,7 +550,6 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
switch ((jj_ntk == -1) ? Jj_ntk() : jj_ntk)
{
case RegexpToken.CARAT:
- ;
break;
default:
jj_la1[9] = jj_gen;
@@ -580,7 +573,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
}
q.Weight = (f * q.Weight); /* left associative, fwiw */
}
- label_8: ;
+ label_8: {/* LUCENENET: intentionally blank */}
}
private bool Jj_2_1(int xla)
@@ -699,7 +692,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
throw GenerateParseException();
}
- // LUCENENET: It is no longer good practice to use binary serialization.
+ // LUCENENET: It is no longer good practice to use binary serialization.
// See: https://github.com/dotnet/corefx/issues/23584#issuecomment-325724568
#if FEATURE_SERIALIZABLE_EXCEPTIONS
[Serializable]
@@ -818,7 +811,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
jj_expentries.Add(jj_expentry);
goto jj_entries_loop_break;
}
- jj_entries_loop_continue: ;
+ jj_entries_loop_continue: {/* LUCENENET: intentionally blank */}
}
jj_entries_loop_break:
if (pos != 0) jj_lasttokens[(jj_endpos = pos) - 1] = kind;
diff --git a/src/Lucene.Net.QueryParser/Surround/Parser/QueryParserTokenManager.cs b/src/Lucene.Net.QueryParser/Surround/Parser/QueryParserTokenManager.cs
index 164fc8dbb..c4999c4bc 100644
--- a/src/Lucene.Net.QueryParser/Surround/Parser/QueryParserTokenManager.cs
+++ b/src/Lucene.Net.QueryParser/Surround/Parser/QueryParserTokenManager.cs
@@ -484,7 +484,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
else if (m_curChar < 128)
{
//long l = 1L << (curChar & 077);
- ulong l = (ulong)(1L << (m_curChar & 63));
+ ulong l = (ulong)(1L << (m_curChar & 63));
do
{
switch (jjstateSet[--i])
@@ -524,8 +524,8 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
}
}
internal static readonly int[] jjnextStates = {
- 32, 33, 34, 35, 37, 24, 27, 28, 20, 17, 21, 18, 27, 28, 30, 24,
- 25, 0, 1,
+ 32, 33, 34, 35, 37, 24, 27, 28, 20, 17, 21, 18, 27, 28, 30, 24,
+ 25, 0, 1,
};
private static bool JjCanMove_0(int hiByte, int i1, int i2, ulong l1, ulong l2)
{
@@ -542,14 +542,14 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
////** Token literal values. */
////public static readonly string[] jjstrLiteralImages = {
- //// "", null, null, null, null, null, null, null, null, null, null, null, null,
- //// "\50", "\51", "\54", "\72", "\136", null, null, null, null, null, null
+ //// "", null, null, null, null, null, null, null, null, null, null, null, null,
+ //// "\50", "\51", "\54", "\72", "\136", null, null, null, null, null, null
////};
/// <summary>Token literal values.</summary>
public static readonly string[] jjstrLiteralImages = {
- "", null, null, null, null, null, null, null, null, null, null, null, null,
- "\x0028" /*"\50"*/, "\x0029" /*"\51"*/, "\x002C" /*"\54"*/, "\x003A" /*"\72"*/, "\x005E" /*"\136"*/, null, null, null, null, null, null
+ "", null, null, null, null, null, null, null, null, null, null, null, null,
+ "\x0028" /*"\50"*/, "\x0029" /*"\51"*/, "\x002C" /*"\54"*/, "\x003A" /*"\72"*/, "\x005E" /*"\136"*/, null, null, null, null, null, null
};
/// <summary>Lexer state names.</summary>
@@ -560,13 +560,13 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
/// <summary>Lex State array.</summary>
public static readonly int[] jjnewLexState = {
- -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, 0, -1, -1, -1, -1, -1, 1,
+ -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, 0, -1, -1, -1, -1, -1, 1,
};
internal static readonly long[] jjtoToken = {
- 0xffff01L,
+ 0xffff01L,
};
internal static readonly long[] jjtoSkip = {
- 0x80L,
+ 0x80L,
};
protected ICharStream m_input_stream;
private readonly uint[] jjrounds = new uint[38];
@@ -723,7 +723,7 @@ namespace Lucene.Net.QueryParsers.Surround.Parser
error_after = curPos <= 1 ? "" : m_input_stream.Image;
}
throw new TokenMgrError(EOFSeen, curLexState, error_line, error_column, error_after, m_curChar, TokenMgrError.LEXICAL_ERROR);
- EOFLoop: ;
+ EOFLoop: {/* LUCENENET: intentionally blank */}
}
}
diff --git a/src/Lucene.Net.Spatial/Prefix/AbstractVisitingPrefixTreeFilter.cs b/src/Lucene.Net.Spatial/Prefix/AbstractVisitingPrefixTreeFilter.cs
index f403253ac..4f2a53adb 100644
--- a/src/Lucene.Net.Spatial/Prefix/AbstractVisitingPrefixTreeFilter.cs
+++ b/src/Lucene.Net.Spatial/Prefix/AbstractVisitingPrefixTreeFilter.cs
@@ -46,7 +46,7 @@ namespace Lucene.Net.Spatial.Prefix
public int PrefixGridScanLevel { get; }//at least one less than grid.getMaxLevels()
- protected AbstractVisitingPrefixTreeFilter(IShape queryShape, string fieldName, SpatialPrefixTree grid,
+ protected AbstractVisitingPrefixTreeFilter(IShape queryShape, string fieldName, SpatialPrefixTree grid,
int detailLevel, int prefixGridScanLevel) // LUCENENET: CA1012: Abstract types should not have constructors (marked protected)
: base(queryShape, fieldName, grid, detailLevel)
{
@@ -145,7 +145,7 @@ namespace Lucene.Net.Spatial.Prefix
return null;// all done
}
thisTerm = m_termsEnum.Term;
-
+
curVNode = new VNode(null);
curVNode.Reset(m_filter.m_grid.WorldCell);
@@ -238,10 +238,9 @@ namespace Lucene.Net.Spatial.Prefix
AddIntersectingChildren();
}
}
- ;
}//main loop
main_break: { }
-
+
return Finish();
}
@@ -279,7 +278,7 @@ namespace Lucene.Net.Spatial.Prefix
thisTerm = m_termsEnum.Term;
}
}
-
+
//Decide whether to continue to divide & conquer, or whether it's time to
// scan through terms beneath this cell.
// Scanning is a performance optimization trade-off.
diff --git a/src/Lucene.Net.Suggest/Spell/SpellChecker.cs b/src/Lucene.Net.Suggest/Spell/SpellChecker.cs
index dc86f149e..ae4b4232b 100644
--- a/src/Lucene.Net.Suggest/Spell/SpellChecker.cs
+++ b/src/Lucene.Net.Suggest/Spell/SpellChecker.cs
@@ -33,9 +33,9 @@ namespace Lucene.Net.Search.Spell
/// Spell Checker class (Main class) <br/>
/// (initially inspired by the David Spencer code).
/// </para>
- ///
+ ///
/// <para>Example Usage (C#):
- ///
+ ///
/// <code>
/// SpellChecker spellchecker = new SpellChecker(spellIndexDirectory);
/// // To index a field of a user index:
@@ -44,7 +44,7 @@ namespace Lucene.Net.Search.Spell
/// spellchecker.IndexDictionary(new PlainTextDictionary(new FileInfo("myfile.txt")));
/// string[] suggestions = spellchecker.SuggestSimilar("misspelt", 5);
/// </code>
- ///
+ ///
/// </para>
/// </summary>
public class SpellChecker : IDisposable
@@ -186,8 +186,8 @@ namespace Lucene.Net.Search.Spell
}
/// <summary>
- /// Gets or sets the accuracy (minimum score) to be used, unless overridden in
- /// <see cref="SuggestSimilar(string, int, IndexReader, string, SuggestMode, float)"/>,
+ /// Gets or sets the accuracy (minimum score) to be used, unless overridden in
+ /// <see cref="SuggestSimilar(string, int, IndexReader, string, SuggestMode, float)"/>,
/// to decide whether a suggestion is included or not.
/// Sets the accuracy 0 < minScore < 1; default <see cref="DEFAULT_ACCURACY"/>
/// </summary>
@@ -256,7 +256,7 @@ namespace Lucene.Net.Search.Spell
/// <summary>
/// Calls <see cref="SuggestSimilar(string, int, IndexReader, string, SuggestMode, float)"/>
/// SuggestSimilar(word, numSug, ir, suggestMode, field, this.accuracy)
- ///
+ ///
/// </summary>
[MethodImpl(MethodImplOptions.AggressiveInlining)]
public virtual string[] SuggestSimilar(string word, int numSug, IndexReader ir, string field, SuggestMode suggestMode)
@@ -282,7 +282,7 @@ namespace Lucene.Net.Search.Spell
/// <param name="ir"> the indexReader of the user index (can be null see field param) </param>
/// <param name="field"> the field of the user index: if field is not null, the suggested
/// words are restricted to the words present in this field. </param>
- /// <param name="suggestMode">
+ /// <param name="suggestMode">
/// (NOTE: if indexReader==null and/or field==null, then this is overridden with SuggestMode.SUGGEST_ALWAYS) </param>
/// <param name="accuracy"> The minimum score a suggestion must have in order to qualify for inclusion in the results </param>
/// <exception cref="IOException"> if the underlying index throws an <see cref="IOException"/> </exception>
@@ -551,8 +551,7 @@ namespace Lucene.Net.Search.Spell
// ok index the word
var doc = CreateDocument(word, GetMin(len), GetMax(len));
writer.AddDocument(doc);
- termsContinue:
- ;
+ termsContinue: {/* LUCENENET: intentionally blank */}
}
}
finally
@@ -667,7 +666,7 @@ namespace Lucene.Net.Search.Spell
[MethodImpl(MethodImplOptions.AggressiveInlining)]
private static void ReleaseSearcher(IndexSearcher aSearcher) // LUCENENET: CA1822: Mark members as static
{
- // don't check if open - always decRef
+ // don't check if open - always decRef
// don't decrement the private searcher - could have been swapped
aSearcher.IndexReader.DecRef();
}
diff --git a/src/Lucene.Net.Suggest/Suggest/Analyzing/FreeTextSuggester.cs b/src/Lucene.Net.Suggest/Suggest/Analyzing/FreeTextSuggester.cs
index a0b6f57f1..80e33a242 100644
--- a/src/Lucene.Net.Suggest/Suggest/Analyzing/FreeTextSuggester.cs
+++ b/src/Lucene.Net.Suggest/Suggest/Analyzing/FreeTextSuggester.cs
@@ -39,7 +39,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
// TODO
// - test w/ syns
- // - add pruning of low-freq ngrams?
+ // - add pruning of low-freq ngrams?
/// <summary>
/// Builds an ngram model from the text sent to <see cref="Build(IInputEnumerator, double)"/>
@@ -47,30 +47,30 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// the request sent to <see cref="DoLookup(string, IEnumerable{BytesRef}, bool, int)"/>. This tries to
/// handle the "long tail" of suggestions for when the
/// incoming query is a never before seen query string.
- ///
+ ///
/// <para>Likely this suggester would only be used as a
/// fallback, when the primary suggester fails to find
/// any suggestions.
- ///
+ ///
/// </para>
/// <para>Note that the weight for each suggestion is unused,
/// and the suggestions are the analyzed forms (so your
/// analysis process should normally be very "light").
- ///
+ ///
/// </para>
/// <para>This uses the stupid backoff language model to smooth
- /// scores across ngram models; see
+ /// scores across ngram models; see
/// <a href="http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.76.1126">
/// "Large language models in machine translation"</a> for details.
- ///
+ ///
/// </para>
/// <para> From <see cref="DoLookup(string, IEnumerable{BytesRef}, bool, int)"/>, the key of each result is the
/// ngram token; the value is <see cref="long.MaxValue"/> * score (fixed
/// point, cast to long). Divide by <see cref="long.MaxValue"/> to get
/// the score back, which ranges from 0.0 to 1.0.
- ///
+ ///
/// <c>onlyMorePopular</c> is unused.
- ///
+ ///
/// @lucene.experimental
/// </para>
/// </summary>
@@ -100,7 +100,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// lookup, this means that if a given trigram did not
/// occur, and we backoff to the bigram, the overall score
/// will be 0.4 times what the bigram model would have
- /// assigned.
+ /// assigned.
/// </summary>
public const double ALPHA = 0.4;
@@ -134,13 +134,13 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// <summary>
/// The default character used to join multiple tokens
/// into a single ngram token. The input tokens produced
- /// by the analyzer must not contain this character.
+ /// by the analyzer must not contain this character.
/// </summary>
public const byte DEFAULT_SEPARATOR = 0x1e;
/// <summary>
/// Instantiate, using the provided analyzer for both
- /// indexing and lookup, using bigram model by default.
+ /// indexing and lookup, using bigram model by default.
/// </summary>
public FreeTextSuggester(Analyzer analyzer)
: this(analyzer, analyzer, DEFAULT_GRAMS)
@@ -149,7 +149,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// <summary>
/// Instantiate, using the provided indexing and lookup
- /// analyzers, using bigram model by default.
+ /// analyzers, using bigram model by default.
/// </summary>
public FreeTextSuggester(Analyzer indexAnalyzer, Analyzer queryAnalyzer)
: this(indexAnalyzer, queryAnalyzer, DEFAULT_GRAMS)
@@ -159,7 +159,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// <summary>
/// Instantiate, using the provided indexing and lookup
/// analyzers, with the specified model (2
- /// = bigram, 3 = trigram, etc.).
+ /// = bigram, 3 = trigram, etc.).
/// </summary>
public FreeTextSuggester(Analyzer indexAnalyzer, Analyzer queryAnalyzer, int grams)
: this(indexAnalyzer, queryAnalyzer, grams, DEFAULT_SEPARATOR)
@@ -174,7 +174,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// tokens into a single ngram token; it must be an ascii
/// (7-bit-clean) byte. No input tokens should have this
/// byte, otherwise <see cref="ArgumentException"/> is
- /// thrown.
+ /// thrown.
/// </summary>
public FreeTextSuggester(Analyzer indexAnalyzer, Analyzer queryAnalyzer, int grams, byte separator)
{
@@ -679,7 +679,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
// reject up to seen.size() paths in acceptResult():
Util.Fst.Util.TopNSearcher<Int64> searcher = new TopNSearcherAnonymousClass(this, fst, num, num + seen.Count, weightComparer, seen, finalLastToken);
- // since this search is initialized with a single start node
+ // since this search is initialized with a single start node
// it is okay to start with an empty input path here
searcher.AddStartPaths(arc, prefixOutput, true, new Int32sRef());
@@ -726,13 +726,13 @@ namespace Lucene.Net.Search.Suggest.Analyzing
spare.Grow(token.Length);
UnicodeUtil.UTF8toUTF16(token, spare);
LookupResult result = new LookupResult(spare.ToString(),
- // LUCENENET NOTE: We need to calculate this as decimal because when using double it can sometimes
+ // LUCENENET NOTE: We need to calculate this as decimal because when using double it can sometimes
// return numbers that are greater than long.MaxValue, which results in a negative long number.
(long)(long.MaxValue * (decimal)backoff * ((decimal)DecodeWeight(completion.Output)) / contextCount));
results.Add(result);
if (Debugging.AssertsEnabled) Debugging.Assert(results.Count == seen.Count);
//System.out.println(" add result=" + result);
- nextCompletionContinue:;
+ nextCompletionContinue: {/* LUCENENET: intentionally blank */}
}
backoff *= ALPHA;
}
diff --git a/src/Lucene.Net/Codecs/BlockTreeTermsReader.cs b/src/Lucene.Net/Codecs/BlockTreeTermsReader.cs
index 3fb848c52..a4d00156a 100644
--- a/src/Lucene.Net/Codecs/BlockTreeTermsReader.cs
+++ b/src/Lucene.Net/Codecs/BlockTreeTermsReader.cs
@@ -1444,7 +1444,7 @@ namespace Lucene.Net.Codecs
{
//System.out.println(" no s=" + state);
}
- nextTermContinue:;
+ nextTermContinue: {/* LUCENENET: intentionally blank */}
}
//nextTermBreak:;
}
@@ -3223,7 +3223,7 @@ namespace Lucene.Net.Codecs
return SeekStatus.FOUND;
}
}
- nextTermContinue: ;
+ nextTermContinue: {/* LUCENENET: intentionally blank */}
}
nextTermBreak:
@@ -3379,7 +3379,7 @@ namespace Lucene.Net.Codecs
return SeekStatus.FOUND;
}
}
- nextTermContinue: ;
+ nextTermContinue: {/* LUCENENET: intentionally blank */}
}
nextTermBreak:
diff --git a/src/Lucene.Net/Codecs/Compressing/LZ4.cs b/src/Lucene.Net/Codecs/Compressing/LZ4.cs
index 55f1dbea5..9beb3aa9f 100644
--- a/src/Lucene.Net/Codecs/Compressing/LZ4.cs
+++ b/src/Lucene.Net/Codecs/Compressing/LZ4.cs
@@ -297,7 +297,7 @@ namespace Lucene.Net.Codecs.Compressing
anchor = off;
//mainContinue: ; // LUCENENET NOTE: Not Referenced
}
- mainBreak: ;
+ mainBreak: {/* LUCENENET: intentionally blank */}
}
// last literals
@@ -641,15 +641,15 @@ namespace Lucene.Net.Codecs.Compressing
CopyTo(match3, match2);
// goto search3Continue; // LUCENENET: Removed redundant jump statements. https://rules.sonarsource.com/csharp/RSPEC-3626
- search3Continue: ;
+ search3Continue: {/* LUCENENET: intentionally blank */}
}
//search3Break: ; // LUCENENET NOTE: Unreachable
- search2Continue: ;
+ search2Continue: {/* LUCENENET: intentionally blank */}
}
//search2Break: ; // LUCENENET NOTE: Not referenced
- mainContinue: ;
+ mainContinue: {/* LUCENENET: intentionally blank */}
}
//mainBreak: // LUCENENET NOTE: Not referenced
diff --git a/src/Lucene.Net/Search/FieldCacheRangeFilter.cs b/src/Lucene.Net/Search/FieldCacheRangeFilter.cs
index 34f27edb4..40887fdd4 100644
--- a/src/Lucene.Net/Search/FieldCacheRangeFilter.cs
+++ b/src/Lucene.Net/Search/FieldCacheRangeFilter.cs
@@ -185,7 +185,7 @@ namespace Lucene.Net.Search
if (inclusiveUpperPoint < 0 || inclusiveLowerPoint > inclusiveUpperPoint)
{
- return null; ;
+ return null;
}
if (Debugging.AssertsEnabled) Debugging.Assert(inclusiveLowerPoint >= 0 && inclusiveUpperPoint >= 0);
@@ -769,7 +769,7 @@ namespace Lucene.Net.Search
}
return true;
}
-
+
// LUCENENET specific - override this method to eliminate boxing on value types
protected virtual bool Equals(T objA, T objB)
{
diff --git a/src/Lucene.Net/Search/SloppyPhraseScorer.cs b/src/Lucene.Net/Search/SloppyPhraseScorer.cs
index 917097406..0bc0571b4 100644
--- a/src/Lucene.Net/Search/SloppyPhraseScorer.cs
+++ b/src/Lucene.Net/Search/SloppyPhraseScorer.cs
@@ -251,7 +251,7 @@ namespace Lucene.Net.Search
/// <item><description>no repetitions: <b>"ho my"~2</b></description></item>
/// <item><description>>repetitions: <b>"ho my my"~2</b></description></item>
/// <item><description>repetitions: <b>"my ho my"~2</b></description></item>
- /// </list>
+ /// </list>
/// </summary>
/// <returns> <c>false</c> if PPs are exhausted (and so current doc will not be a match) </returns>
private bool InitPhrasePositions()
@@ -449,7 +449,7 @@ namespace Lucene.Net.Search
}
}
}
-
+
/// <summary>
/// Detect repetition groups. Done once - for first doc. </summary>
private IList<IList<PhrasePositions>> GatherRptGroups(JCG.LinkedDictionary<Term, int> rptTerms)
@@ -542,7 +542,6 @@ namespace Lucene.Net.Search
{
foreach (Term t in pp.terms)
{
- ;
int cnt = !tcnt.TryGetValue(t, out int cnt0) ? 1 : 1 + cnt0;
tcnt[t] = cnt;
if (cnt == 2)
diff --git a/src/Lucene.Net/Store/CompoundFileWriter.cs b/src/Lucene.Net/Store/CompoundFileWriter.cs
index a1e53cbf8..61b229533 100644
--- a/src/Lucene.Net/Store/CompoundFileWriter.cs
+++ b/src/Lucene.Net/Store/CompoundFileWriter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Store
/// <summary>
/// Combines multiple files into a single compound file.
/// <para/>
- /// @lucene.internal
+ /// @lucene.internal
/// </summary>
/// <seealso cref="CompoundFileDirectory"/>
internal sealed class CompoundFileWriter : IDisposable
@@ -311,7 +311,7 @@ namespace Lucene.Net.Store
while (pendingEntries.Count > 0)
{
FileEntry entry = pendingEntries.First.Value;
- pendingEntries.Remove(entry); ;
+ pendingEntries.Remove(entry);
CopyFileEntry(GetOutput(), entry);
entries[entry.File] = entry;
}
diff --git a/src/Lucene.Net/Util/WAH8DocIdSet.cs b/src/Lucene.Net/Util/WAH8DocIdSet.cs
index 45fbbda1d..8bb85057b 100644
--- a/src/Lucene.Net/Util/WAH8DocIdSet.cs
+++ b/src/Lucene.Net/Util/WAH8DocIdSet.cs
@@ -89,7 +89,7 @@ namespace Lucene.Net.Util
/// <summary>
/// Default index interval. </summary>
public const int DEFAULT_INDEX_INTERVAL = 24;
-
+
private static readonly MonotonicAppendingInt64Buffer SINGLE_ZERO_BUFFER = LoadSingleZeroBuffer();
// LUCENENET specific - optimized empty array creation
private static readonly WAH8DocIdSet EMPTY = new WAH8DocIdSet(Arrays.Empty<byte>(), 0, 1, SINGLE_ZERO_BUFFER, SINGLE_ZERO_BUFFER);
@@ -175,7 +175,7 @@ namespace Lucene.Net.Util
if (Debugging.AssertsEnabled) Debugging.Assert(word != 0);
builder.AddWord(wordNum, word);
++wordNum;
- mainContinue:;
+ mainContinue: {/* LUCENENET: intentionally blank */}
}
//mainBreak:
return builder.Build();