You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucenenet.apache.org by ni...@apache.org on 2017/06/01 22:48:58 UTC
[2/9] lucenenet git commit: SWEEP: Changed to
in documentation comments
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Pt/RSLPStemmerBase.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Pt/RSLPStemmerBase.cs b/src/Lucene.Net.Analysis.Common/Analysis/Pt/RSLPStemmerBase.cs
index 04b6c1d..e36aee8 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Pt/RSLPStemmerBase.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Pt/RSLPStemmerBase.cs
@@ -44,14 +44,14 @@ namespace Lucene.Net.Analysis.Pt
/// <code>{ "suffix", N, "replacement", { "exception1", "exception2", ...}}</code>
/// where:
/// <list type="bullet">
- /// <item><c>suffix</c> is the suffix to be removed (such as "inho").</item>
- /// <item><c>N</c> is the min stem size, where stem is defined as the candidate stem
- /// after removing the suffix (but before appending the replacement!)</item>
- /// <item><c>replacement</c> is an optimal string to append after removing the suffix.
- /// This can be the empty string.</item>
- /// <item><c>exceptions</c> is an optional list of exceptions, patterns that should
+ /// <item><description><c>suffix</c> is the suffix to be removed (such as "inho").</description></item>
+ /// <item><description><c>N</c> is the min stem size, where stem is defined as the candidate stem
+ /// after removing the suffix (but before appending the replacement!)</description></item>
+ /// <item><description><c>replacement</c> is an optimal string to append after removing the suffix.
+ /// This can be the empty string.</description></item>
+ /// <item><description><c>exceptions</c> is an optional list of exceptions, patterns that should
/// not be stemmed. These patterns can be specified as whole word or suffix (ends-with)
- /// patterns, depending upon the exceptions format flag in the step header.</item>
+ /// patterns, depending upon the exceptions format flag in the step header.</description></item>
/// </list>
/// </para>
/// <para>
@@ -61,17 +61,17 @@ namespace Lucene.Net.Analysis.Pt
/// </blockquote>
/// where:
/// <list type="bullet">
- /// <item><c>name</c> is a name for the step (such as "Plural").</item>
- /// <item><c>N</c> is the min word size. Words that are less than this length bypass
+ /// <item><description><c>name</c> is a name for the step (such as "Plural").</description></item>
+ /// <item><description><c>N</c> is the min word size. Words that are less than this length bypass
/// the step completely, as an optimization. Note: N can be zero, in this case this
/// implementation will automatically calculate the appropriate value from the underlying
- /// rules.</item>
- /// <item><c>B</c> is a "boolean" flag specifying how exceptions in the rules are matched.
+ /// rules.</description></item>
+ /// <item><description><c>B</c> is a "boolean" flag specifying how exceptions in the rules are matched.
/// A value of 1 indicates whole-word pattern matching, a value of 0 indicates that
- /// exceptions are actually suffixes and should be matched with ends-with.</item>
- /// <item><c>conds</c> are an optional list of conditions to enter the step at all. If
+ /// exceptions are actually suffixes and should be matched with ends-with.</description></item>
+ /// <item><description><c>conds</c> are an optional list of conditions to enter the step at all. If
/// the list is non-empty, then a word must end with one of these conditions or it will
- /// bypass the step completely as an optimization.</item>
+ /// bypass the step completely as an optimization.</description></item>
/// </list>
/// </para>
/// <a href="http://www.inf.ufrgs.br/~viviane/rslp/index.htm">RSLP description</a>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Reverse/ReverseStringFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Reverse/ReverseStringFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Reverse/ReverseStringFilter.cs
index 4b48902..4440167 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Reverse/ReverseStringFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Reverse/ReverseStringFilter.cs
@@ -33,7 +33,7 @@ namespace Lucene.Net.Analysis.Reverse
/// compatibility when creating <see cref="ReverseStringFilter"/>, or when using any of
/// its static methods:
/// <list type="bullet">
- /// <item> As of 3.1, supplementary characters are handled correctly</item>
+ /// <item><description> As of 3.1, supplementary characters are handled correctly</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Ru/RussianAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Ru/RussianAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Ru/RussianAnalyzer.cs
index ad8e0ea..bb086a7 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Ru/RussianAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Ru/RussianAnalyzer.cs
@@ -38,8 +38,8 @@ namespace Lucene.Net.Analysis.Ru
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="RussianAnalyzer"/>:
/// <list type="bullet">
- /// <item> As of 3.1, <see cref="StandardTokenizer"/> is used, Snowball stemming is done with
- /// <see cref="SnowballFilter"/>, and Snowball stopwords are used by default.</item>
+ /// <item><description> As of 3.1, <see cref="StandardTokenizer"/> is used, Snowball stemming is done with
+ /// <see cref="SnowballFilter"/>, and Snowball stopwords are used by default.</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballAnalyzer.cs
index bcaa1d6..a6d2be2 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballAnalyzer.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Snowball
/// <para><b>NOTE</b>: This class uses the same <see cref="LuceneVersion"/>
/// dependent settings as <see cref="StandardAnalyzer"/>, with the following addition:
/// <list type="bullet">
- /// <item> As of 3.1, uses <see cref="TurkishLowerCaseFilter"/> for Turkish language.</item>
+ /// <item><description> As of 3.1, uses <see cref="TurkishLowerCaseFilter"/> for Turkish language.</description></item>
/// </list>
/// </para> </summary>
/// @deprecated (3.1) Use the language-specific analyzer in modules/analysis instead.
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballFilter.cs
index d9f8672..8ac2021 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Snowball/SnowballFilter.cs
@@ -28,8 +28,8 @@ namespace Lucene.Net.Analysis.Snowball
/// Available stemmers are listed in Lucene.Net.Tartarus.Snowball.Ext.
/// <para><b>NOTE</b>: <see cref="SnowballFilter"/> expects lowercased text.
/// <list type="bullet">
- /// <item>For the Turkish language, see <see cref="Tr.TurkishLowerCaseFilter"/>.</item>
- /// <item>For other languages, see <see cref="Core.LowerCaseFilter"/>.</item>
+ /// <item><description>For the Turkish language, see <see cref="Tr.TurkishLowerCaseFilter"/>.</description></item>
+ /// <item><description>For other languages, see <see cref="Core.LowerCaseFilter"/>.</description></item>
/// </list>
/// </para>
///
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicAnalyzer.cs
index 888431b..f5b42e0 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicAnalyzer.cs
@@ -30,12 +30,12 @@ namespace Lucene.Net.Analysis.Standard
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="ClassicAnalyzer"/>:
/// <list type="bullet">
- /// <item> As of 3.1, <see cref="StopFilter"/> correctly handles Unicode 4.0
- /// supplementary characters in stopwords</item>
- /// <item> As of 2.9, <see cref="StopFilter"/> preserves position
- /// increments</item>
- /// <item> As of 2.4, <see cref="Token"/>s incorrectly identified as acronyms
- /// are corrected (see <a href="https://issues.apache.org/jira/browse/LUCENE-1068">LUCENE-1068</a>)</item>
+ /// <item><description> As of 3.1, <see cref="StopFilter"/> correctly handles Unicode 4.0
+ /// supplementary characters in stopwords</description></item>
+ /// <item><description> As of 2.9, <see cref="StopFilter"/> preserves position
+ /// increments</description></item>
+ /// <item><description> As of 2.4, <see cref="Token"/>s incorrectly identified as acronyms
+ /// are corrected (see <a href="https://issues.apache.org/jira/browse/LUCENE-1068">LUCENE-1068</a>)</description></item>
/// </list>
///
/// <see cref="ClassicAnalyzer"/> was named <see cref="StandardAnalyzer"/> in Lucene versions prior to 3.1.
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicTokenizer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicTokenizer.cs
index 111e23f..1bd65af 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicTokenizer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/ClassicTokenizer.cs
@@ -27,11 +27,11 @@ namespace Lucene.Net.Analysis.Standard
/// <para> This should be a good tokenizer for most European-language documents:
///
/// <list type="bullet">
- /// <item>Splits words at punctuation characters, removing punctuation. However, a
- /// dot that's not followed by whitespace is considered part of a token.</item>
- /// <item>Splits words at hyphens, unless there's a number in the token, in which case
- /// the whole token is interpreted as a product number and is not split.</item>
- /// <item>Recognizes email addresses and internet hostnames as one token.</item>
+ /// <item><description>Splits words at punctuation characters, removing punctuation. However, a
+ /// dot that's not followed by whitespace is considered part of a token.</description></item>
+ /// <item><description>Splits words at hyphens, unless there's a number in the token, in which case
+ /// the whole token is interpreted as a product number and is not split.</description></item>
+ /// <item><description>Recognizes email addresses and internet hostnames as one token.</description></item>
/// </list>
///
/// </para>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardAnalyzer.cs
index d7f8515..ca6c60c 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardAnalyzer.cs
@@ -30,17 +30,17 @@ namespace Lucene.Net.Analysis.Standard
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="StandardAnalyzer"/>:
/// <list type="bullet">
- /// <item> As of 3.4, Hiragana and Han characters are no longer wrongly split
+ /// <item><description> As of 3.4, Hiragana and Han characters are no longer wrongly split
/// from their combining characters. If you use a previous version number,
- /// you get the exact broken behavior for backwards compatibility.</item>
- /// <item> As of 3.1, <see cref="StandardTokenizer"/> implements Unicode text segmentation,
+ /// you get the exact broken behavior for backwards compatibility.</description></item>
+ /// <item><description> As of 3.1, <see cref="StandardTokenizer"/> implements Unicode text segmentation,
/// and <see cref="StopFilter"/> correctly handles Unicode 4.0 supplementary characters
/// in stopwords. <see cref="ClassicTokenizer"/> and <see cref="ClassicAnalyzer"/>
/// are the pre-3.1 implementations of <see cref="StandardTokenizer"/> and
- /// <see cref="StandardAnalyzer"/>.</item>
- /// <item> As of 2.9, <see cref="StopFilter"/> preserves position increments</item>
- /// <item> As of 2.4, <see cref="Token"/>s incorrectly identified as acronyms
- /// are corrected (see <a href="https://issues.apache.org/jira/browse/LUCENE-1068">LUCENE-1068</a>)</item>
+ /// <see cref="StandardAnalyzer"/>.</description></item>
+ /// <item><description> As of 2.9, <see cref="StopFilter"/> preserves position increments</description></item>
+ /// <item><description> As of 2.4, <see cref="Token"/>s incorrectly identified as acronyms
+ /// are corrected (see <a href="https://issues.apache.org/jira/browse/LUCENE-1068">LUCENE-1068</a>)</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizer.cs
index 3fa7bb8..5d89a29 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizer.cs
@@ -41,12 +41,12 @@ namespace Lucene.Net.Analysis.Standard
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="StandardTokenizer"/>:
/// <list type="bullet">
- /// <item> As of 3.4, Hiragana and Han characters are no longer wrongly split
+ /// <item><description> As of 3.4, Hiragana and Han characters are no longer wrongly split
/// from their combining characters. If you use a previous version number,
- /// you get the exact broken behavior for backwards compatibility.</item>
- /// <item> As of 3.1, StandardTokenizer implements Unicode text segmentation.
+ /// you get the exact broken behavior for backwards compatibility.</description></item>
+ /// <item><description> As of 3.1, StandardTokenizer implements Unicode text segmentation.
/// If you use a previous version number, you get the exact behavior of
- /// <see cref="ClassicTokenizer"/> for backwards compatibility.</item>
+ /// <see cref="ClassicTokenizer"/> for backwards compatibility.</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizerImpl.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizerImpl.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizerImpl.cs
index aa66336..edfcbb8 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizerImpl.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/StandardTokenizerImpl.cs
@@ -29,14 +29,14 @@ namespace Lucene.Net.Analysis.Standard
/// <para/>
/// Tokens produced are of the following types:
/// <list type="bullet">
- /// <item><ALPHANUM>: A sequence of alphabetic and numeric characters</item>
- /// <item><NUM>: A number</item>
- /// <item><SOUTHEAST_ASIAN>: A sequence of characters from South and Southeast
- /// Asian languages, including Thai, Lao, Myanmar, and Khmer</item>
- /// <item><IDEOGRAPHIC>: A single CJKV ideographic character</item>
- /// <item><HIRAGANA>: A single hiragana character</item>
- /// <item><KATAKANA>: A sequence of katakana characters</item>
- /// <item><HANGUL>: A sequence of Hangul characters</item>
+ /// <item><description><ALPHANUM>: A sequence of alphabetic and numeric characters</description></item>
+ /// <item><description><NUM>: A number</description></item>
+ /// <item><description><SOUTHEAST_ASIAN>: A sequence of characters from South and Southeast
+ /// Asian languages, including Thai, Lao, Myanmar, and Khmer</description></item>
+ /// <item><description><IDEOGRAPHIC>: A single CJKV ideographic character</description></item>
+ /// <item><description><HIRAGANA>: A single hiragana character</description></item>
+ /// <item><description><KATAKANA>: A sequence of katakana characters</description></item>
+ /// <item><description><HANGUL>: A sequence of Hangul characters</description></item>
/// </list>
/// </summary>
public sealed class StandardTokenizerImpl : IStandardTokenizerInterface
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizer.cs
index 292b7bd..31642c6 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizer.cs
@@ -33,21 +33,21 @@ namespace Lucene.Net.Analysis.Standard
/// <para/>
/// Tokens produced are of the following types:
/// <list type="bullet">
- /// <item><ALPHANUM>: A sequence of alphabetic and numeric characters</item>
- /// <item><NUM>: A number</item>
- /// <item><URL>: A URL</item>
- /// <item><EMAIL>: An email address</item>
- /// <item><SOUTHEAST_ASIAN>: A sequence of characters from South and Southeast
- /// Asian languages, including Thai, Lao, Myanmar, and Khmer</item>
- /// <item><IDEOGRAPHIC>: A single CJKV ideographic character</item>
- /// <item><HIRAGANA>: A single hiragana character</item>
+ /// <item><description><ALPHANUM>: A sequence of alphabetic and numeric characters</description></item>
+ /// <item><description><NUM>: A number</description></item>
+ /// <item><description><URL>: A URL</description></item>
+ /// <item><description><EMAIL>: An email address</description></item>
+ /// <item><description><SOUTHEAST_ASIAN>: A sequence of characters from South and Southeast
+ /// Asian languages, including Thai, Lao, Myanmar, and Khmer</description></item>
+ /// <item><description><IDEOGRAPHIC>: A single CJKV ideographic character</description></item>
+ /// <item><description><HIRAGANA>: A single hiragana character</description></item>
/// </list>
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="UAX29URLEmailTokenizer"/>:
/// <list type="bullet">
- /// <item> As of 3.4, Hiragana and Han characters are no longer wrongly split
+ /// <item><description> As of 3.4, Hiragana and Han characters are no longer wrongly split
/// from their combining characters. If you use a previous version number,
- /// you get the exact broken behavior for backwards compatibility.</item>
+ /// you get the exact broken behavior for backwards compatibility.</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizerImpl.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizerImpl.cs b/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizerImpl.cs
index 547a62c..c95f064 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizerImpl.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Standard/UAX29URLEmailTokenizerImpl.cs
@@ -29,16 +29,16 @@ namespace Lucene.Net.Analysis.Standard
/// <para/>
/// Tokens produced are of the following types:
/// <list type="bullet">
- /// <item><ALPHANUM>: A sequence of alphabetic and numeric characters</item>
- /// <item><NUM>: A number</item>
- /// <item><URL>: A URL</item>
- /// <item><EMAIL>: An email address</item>
- /// <item><SOUTHEAST_ASIAN>: A sequence of characters from South and Southeast
- /// Asian languages, including Thai, Lao, Myanmar, and Khmer</item>
- /// <item><IDEOGRAPHIC>: A single CJKV ideographic character</item>
- /// <item><HIRAGANA>: A single hiragana character</item>
- /// <item><KATAKANA>: A sequence of katakana characters</item>
- /// <item><HANGUL>: A sequence of Hangul characters</item>
+ /// <item><description><ALPHANUM>: A sequence of alphabetic and numeric characters</description></item>
+ /// <item><description><NUM>: A number</description></item>
+ /// <item><description><URL>: A URL</description></item>
+ /// <item><description><EMAIL>: An email address</description></item>
+ /// <item><description><SOUTHEAST_ASIAN>: A sequence of characters from South and Southeast
+ /// Asian languages, including Thai, Lao, Myanmar, and Khmer</description></item>
+ /// <item><description><IDEOGRAPHIC>: A single CJKV ideographic character</description></item>
+ /// <item><description><HIRAGANA>: A single hiragana character</description></item>
+ /// <item><description><KATAKANA>: A sequence of katakana characters</description></item>
+ /// <item><description><HANGUL>: A sequence of Hangul characters</description></item>
/// </list>
/// </summary>
public sealed class UAX29URLEmailTokenizerImpl : IStandardTokenizerInterface
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SolrSynonymParser.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SolrSynonymParser.cs b/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SolrSynonymParser.cs
index 7016143..74b969c 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SolrSynonymParser.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SolrSynonymParser.cs
@@ -26,21 +26,21 @@ namespace Lucene.Net.Analysis.Synonym
/// <summary>
/// Parser for the Solr synonyms format.
/// <list type="bullet">
- /// <item> Blank lines and lines starting with '#' are comments.</item>
- /// <item> Explicit mappings match any token sequence on the LHS of "=>"
+ /// <item><description> Blank lines and lines starting with '#' are comments.</description></item>
+ /// <item><description> Explicit mappings match any token sequence on the LHS of "=>"
/// and replace with all alternatives on the RHS. These types of mappings
/// ignore the expand parameter in the constructor.
/// Example:
/// <code>i-pod, i pod => ipod</code>
- /// </item>
- /// <item> Equivalent synonyms may be separated with commas and give
+ /// </description></item>
+ /// <item><description> Equivalent synonyms may be separated with commas and give
/// no explicit mapping. In this case the mapping behavior will
/// be taken from the expand parameter in the constructor. This allows
/// the same synonym file to be used in different synonym handling strategies.
/// Example:
/// <code>ipod, i-pod, i pod</code>
- /// </item>
- /// <item> Multiple synonym mapping entries are merged.
+ /// </description></item>
+ /// <item><description> Multiple synonym mapping entries are merged.
/// Example:
/// <code>
/// foo => foo bar
@@ -48,7 +48,7 @@ namespace Lucene.Net.Analysis.Synonym
/// is equivalent to
/// foo => foo bar, baz
/// </code>
- /// </item>
+ /// </description></item>
/// </list>
/// @lucene.experimental
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SynonymFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SynonymFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SynonymFilterFactory.cs
index 80699e6..d08941c 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SynonymFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Synonym/SynonymFilterFactory.cs
@@ -49,9 +49,9 @@ namespace Lucene.Net.Analysis.Synonym
/// <see cref="SynonymMap.Parser"/> class name. The default is <c>solr</c>.
/// A custom <see cref="SynonymMap.Parser"/> is expected to have a constructor taking:
/// <list type="bullet">
- /// <item><c><see cref="bool"/> dedup</c> - true if duplicates should be ignored, false otherwise</item>
- /// <item><c><see cref="bool"/> expand</c> - true if conflation groups should be expanded, false if they are one-directional</item>
- /// <item><c><see cref="Analyzer"/> analyzer</c> - an analyzer used for each raw synonym</item>
+ /// <item><description><c><see cref="bool"/> dedup</c> - true if duplicates should be ignored, false otherwise</description></item>
+ /// <item><description><c><see cref="bool"/> expand</c> - true if conflation groups should be expanded, false if they are one-directional</description></item>
+ /// <item><description><c><see cref="Analyzer"/> analyzer</c> - an analyzer used for each raw synonym</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Th/ThaiAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Th/ThaiAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Th/ThaiAnalyzer.cs
index 0885069..5b84fde 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Th/ThaiAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Th/ThaiAnalyzer.cs
@@ -30,7 +30,7 @@ namespace Lucene.Net.Analysis.Th
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="ThaiAnalyzer"/>:
/// <list type="bullet">
- /// <item> As of 3.6, a set of Thai stopwords is used by default</item>
+ /// <item><description> As of 3.6, a set of Thai stopwords is used by default</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Util/AbstractAnalysisFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Util/AbstractAnalysisFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/Util/AbstractAnalysisFactory.cs
index a3ec443..6761686 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Util/AbstractAnalysisFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Util/AbstractAnalysisFactory.cs
@@ -33,10 +33,10 @@ namespace Lucene.Net.Analysis.Util
/// <para>
/// The typical lifecycle for a factory consumer is:
/// <list type="bullet">
- /// <item>Create factory via its constructor (or via XXXFactory.ForName)</item>
- /// <item>(Optional) If the factory uses resources such as files,
- /// <see cref="IResourceLoaderAware.Inform(IResourceLoader)"/> is called to initialize those resources.</item>
- /// <item>Consumer calls create() to obtain instances.</item>
+ /// <item><description>Create factory via its constructor (or via XXXFactory.ForName)</description></item>
+ /// <item><description>(Optional) If the factory uses resources such as files,
+ /// <see cref="IResourceLoaderAware.Inform(IResourceLoader)"/> is called to initialize those resources.</description></item>
+ /// <item><description>Consumer calls create() to obtain instances.</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Util/CharArrayMap.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Util/CharArrayMap.cs b/src/Lucene.Net.Analysis.Common/Analysis/Util/CharArrayMap.cs
index e006ea5..447fb98 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Util/CharArrayMap.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Util/CharArrayMap.cs
@@ -41,8 +41,8 @@ namespace Lucene.Net.Analysis.Util
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="CharArrayMap"/>:
/// <list type="bullet">
- /// <item> As of 3.1, supplementary characters are
- /// properly lowercased.</item>
+ /// <item><description> As of 3.1, supplementary characters are
+ /// properly lowercased.</description></item>
/// </list>
/// Before 3.1 supplementary characters could not be
/// lowercased correctly due to the lack of Unicode 4
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Util/CharTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Util/CharTokenizer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Util/CharTokenizer.cs
index 9ef33c4..e3ba728 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Util/CharTokenizer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Util/CharTokenizer.cs
@@ -29,9 +29,9 @@ namespace Lucene.Net.Analysis.Util
/// You must specify the required <see cref="LuceneVersion"/> compatibility
/// when creating <see cref="CharTokenizer"/>:
/// <list type="bullet">
- /// <item>As of 3.1, <see cref="CharTokenizer"/> uses an int based API to normalize and
+ /// <item><description>As of 3.1, <see cref="CharTokenizer"/> uses an int based API to normalize and
/// detect token codepoints. See <see cref="IsTokenChar(int)"/> and
- /// <see cref="Normalize(int)"/> for details.</item>
+ /// <see cref="Normalize(int)"/> for details.</description></item>
/// </list>
/// </para>
/// <para>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Analysis/Util/WordlistLoader.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Util/WordlistLoader.cs b/src/Lucene.Net.Analysis.Common/Analysis/Util/WordlistLoader.cs
index 5687823..631879d 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Util/WordlistLoader.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Util/WordlistLoader.cs
@@ -134,9 +134,9 @@ namespace Lucene.Net.Analysis.Util
/// <para>
/// The snowball format is the following:
/// <list type="bullet">
- /// <item>Lines may contain multiple words separated by whitespace.</item>
- /// <item>The comment character is the vertical line (|).</item>
- /// <item>Lines may contain trailing comments.</item>
+ /// <item><description>Lines may contain multiple words separated by whitespace.</description></item>
+ /// <item><description>The comment character is the vertical line (|).</description></item>
+ /// <item><description>Lines may contain trailing comments.</description></item>
/// </list>
/// </para>
/// </summary>
@@ -177,9 +177,9 @@ namespace Lucene.Net.Analysis.Util
/// <para>
/// The snowball format is the following:
/// <list type="bullet">
- /// <item>Lines may contain multiple words separated by whitespace.</item>
- /// <item>The comment character is the vertical line (|).</item>
- /// <item>Lines may contain trailing comments.</item>
+ /// <item><description>Lines may contain multiple words separated by whitespace.</description></item>
+ /// <item><description>The comment character is the vertical line (|).</description></item>
+ /// <item><description>Lines may contain trailing comments.</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Collation/CollationAttributeFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Collation/CollationAttributeFactory.cs b/src/Lucene.Net.Analysis.Common/Collation/CollationAttributeFactory.cs
index bc2fa88..b1d289d 100644
--- a/src/Lucene.Net.Analysis.Common/Collation/CollationAttributeFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Collation/CollationAttributeFactory.cs
@@ -38,16 +38,16 @@ namespace Lucene.Net.Collation
/// same at query time):
/// </para>
/// <list type="number">
- /// <item>JVM vendor</item>
- /// <item>JVM version, including patch version</item>
- /// <item>
+ /// <item><description>JVM vendor</description></item>
+ /// <item><description>JVM version, including patch version</description></item>
+ /// <item><description>
/// The language (and country and variant, if specified) of the Locale
/// used when constructing the collator via
/// <see cref="Collator.Create(System.Globalization.CultureInfo)"/>.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The collation strength used - see <see cref="Collator.Strength"/>
- /// </item>
+ /// </description></item>
/// </list>
/// <para>
/// The <c>ICUCollationAttributeFactory</c> in the analysis-icu package
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
index 92ce4a0..4028f0c 100644
--- a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
@@ -45,16 +45,16 @@ namespace Lucene.Net.Collation
/// same at query time):
/// </para>
/// <list type="number">
- /// <item>JVM vendor</item>
- /// <item>JVM version, including patch version</item>
- /// <item>
+ /// <item><description>JVM vendor</description></item>
+ /// <item><description>JVM version, including patch version</description></item>
+ /// <item><description>
/// The language (and country and variant, if specified) of the Locale
/// used when constructing the collator via
/// <see cref="Collator.Create(System.Globalization.CultureInfo)"/>.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The collation strength used - see <see cref="Collator.Strength"/>
- /// </item>
+ /// </description></item>
/// </list>
/// <para>
/// The <c>ICUCollationKeyAnalyzer</c> in the analysis-icu package
@@ -76,8 +76,8 @@ namespace Lucene.Net.Collation
/// <para>You must specify the required <see cref="LuceneVersion"/>
/// compatibility when creating <see cref="CollationKeyAnalyzer"/>:
/// <list type="bullet">
- /// <item> As of 4.0, Collation Keys are directly encoded as bytes. Previous
- /// versions will encode the bytes with <see cref="IndexableBinaryStringTools"/>.</item>
+ /// <item><description> As of 4.0, Collation Keys are directly encoded as bytes. Previous
+ /// versions will encode the bytes with <see cref="IndexableBinaryStringTools"/>.</description></item>
/// </list>
/// </para>
/// </summary>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilter.cs b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilter.cs
index 4e053d7..d498f84 100644
--- a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilter.cs
@@ -40,16 +40,16 @@ namespace Lucene.Net.Collation
/// same at query time):
/// </para>
/// <list type="number">
- /// <item>JVM vendor</item>
- /// <item>JVM version, including patch version</item>
- /// <item>
+ /// <item><description>JVM vendor</description></item>
+ /// <item><description>JVM version, including patch version</description></item>
+ /// <item><description>
/// The language (and country and variant, if specified) of the Locale
/// used when constructing the collator via
/// <see cref="Collator.Create(System.Globalization.CultureInfo)"/>.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The collation strength used - see <see cref="Collator.Strength"/>
- /// </item>
+ /// </description></item>
/// </list>
/// <para>
/// The <c>ICUCollationKeyFilter</c> in the analysis-icu package
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilterFactory.cs
index 6599f17..5293cb1 100644
--- a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyFilterFactory.cs
@@ -34,26 +34,26 @@ namespace Lucene.Net.Collation
/// <para>
/// This factory can be created in two ways:
/// <list type="bullet">
- /// <item>Based upon a system collator associated with a <see cref="System.Globalization.CultureInfo"/>.</item>
- /// <item>Based upon a tailored ruleset.</item>
+ /// <item><description>Based upon a system collator associated with a <see cref="System.Globalization.CultureInfo"/>.</description></item>
+ /// <item><description>Based upon a tailored ruleset.</description></item>
/// </list>
/// </para>
/// <para>
/// Using a System collator:
/// <list type="bullet">
- /// <item>language: ISO-639 language code (mandatory)</item>
- /// <item>country: ISO-3166 country code (optional)</item>
- /// <item>variant: vendor or browser-specific code (optional)</item>
- /// <item>strength: 'primary','secondary','tertiary', or 'identical' (optional)</item>
- /// <item>decomposition: 'no','canonical', or 'full' (optional)</item>
+ /// <item><description>language: ISO-639 language code (mandatory)</description></item>
+ /// <item><description>country: ISO-3166 country code (optional)</description></item>
+ /// <item><description>variant: vendor or browser-specific code (optional)</description></item>
+ /// <item><description>strength: 'primary','secondary','tertiary', or 'identical' (optional)</description></item>
+ /// <item><description>decomposition: 'no','canonical', or 'full' (optional)</description></item>
/// </list>
/// </para>
/// <para>
/// Using a Tailored ruleset:
/// <list type="bullet">
- /// <item>custom: UTF-8 text file containing rules supported by RuleBasedCollator (mandatory)</item>
- /// <item>strength: 'primary','secondary','tertiary', or 'identical' (optional)</item>
- /// <item>decomposition: 'no','canonical', or 'full' (optional)</item>
+ /// <item><description>custom: UTF-8 text file containing rules supported by RuleBasedCollator (mandatory)</description></item>
+ /// <item><description>strength: 'primary','secondary','tertiary', or 'identical' (optional)</description></item>
+ /// <item><description>decomposition: 'no','canonical', or 'full' (optional)</description></item>
/// </list>
///
/// <code>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Facet/Taxonomy/ParallelTaxonomyArrays.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Facet/Taxonomy/ParallelTaxonomyArrays.cs b/src/Lucene.Net.Facet/Taxonomy/ParallelTaxonomyArrays.cs
index c5db947..6dc81c9 100644
--- a/src/Lucene.Net.Facet/Taxonomy/ParallelTaxonomyArrays.cs
+++ b/src/Lucene.Net.Facet/Taxonomy/ParallelTaxonomyArrays.cs
@@ -23,12 +23,12 @@ namespace Lucene.Net.Facet.Taxonomy
/// <summary>
/// Returns 3 arrays for traversing the taxonomy:
/// <list type="bullet">
- /// <item> <see cref="Parents"/>: <c>Parents[i]</c> denotes the parent of category
- /// ordinal <c>i</c>.</item>
- /// <item> <see cref="Children"/>: <c>Children[i]</c> denotes a child of category ordinal
- /// <c>i</c>.</item>
- /// <item> <see cref="Siblings"/>: <c>Siblings[i]</c> denotes the sibling of category
- /// ordinal <c>i</c>.</item>
+ /// <item><description> <see cref="Parents"/>: <c>Parents[i]</c> denotes the parent of category
+ /// ordinal <c>i</c>.</description></item>
+ /// <item><description> <see cref="Children"/>: <c>Children[i]</c> denotes a child of category ordinal
+ /// <c>i</c>.</description></item>
+ /// <item><description> <see cref="Siblings"/>: <c>Siblings[i]</c> denotes the sibling of category
+ /// ordinal <c>i</c>.</description></item>
/// </list>
///
/// To traverse the taxonomy tree, you typically start with <c>Children[0]</c>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Facet/Taxonomy/TaxonomyReader.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Facet/Taxonomy/TaxonomyReader.cs b/src/Lucene.Net.Facet/Taxonomy/TaxonomyReader.cs
index cbe3742..ca4d6db 100644
--- a/src/Lucene.Net.Facet/Taxonomy/TaxonomyReader.cs
+++ b/src/Lucene.Net.Facet/Taxonomy/TaxonomyReader.cs
@@ -31,15 +31,15 @@ namespace Lucene.Net.Facet.Taxonomy
/// A TaxonomyReader holds a list of categories. Each category has a serial
/// number which we call an "ordinal", and a hierarchical "path" name:
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// The ordinal is an integer that starts at 0 for the first category (which is
/// always the root category), and grows contiguously as more categories are
/// added; Note that once a category is added, it can never be deleted.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The path is a CategoryPath object specifying the category's position in the
/// hierarchy.
- /// </item>
+ /// </description></item>
/// </list>
/// </para>
/// <b>Notes about concurrent access to the taxonomy:</b>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Highlighter/Highlight/TokenSources.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Highlighter/Highlight/TokenSources.cs b/src/Lucene.Net.Highlighter/Highlight/TokenSources.cs
index 05f8e92..c6845e9 100644
--- a/src/Lucene.Net.Highlighter/Highlight/TokenSources.cs
+++ b/src/Lucene.Net.Highlighter/Highlight/TokenSources.cs
@@ -155,35 +155,35 @@ namespace Lucene.Net.Search.Highlight
/// <para/>
/// In my tests the speeds to recreate 1000 token streams using this method are:
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// with TermVector offset only data stored - 420 milliseconds
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// with TermVector offset AND position data stored - 271 milliseconds
/// (nb timings for TermVector with position data are based on a tokenizer with contiguous
/// positions - no overlaps or gaps)
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The cost of not using TermPositionVector to store
/// pre-parsed content and using an analyzer to re-parse the original content:
/// - reanalyzing the original content - 980 milliseconds
- /// </item>
+ /// </description></item>
/// </list>
///
/// The re-analyze timings will typically vary depending on -
/// <list type="number">
- /// <item>
+ /// <item><description>
/// The complexity of the analyzer code (timings above were using a
/// stemmer/lowercaser/stopword combo)
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The number of other fields (Lucene reads ALL fields off the disk
/// when accessing just one document field - can cost dear!)
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// Use of compression on field storage - could be faster due to compression (less disk IO)
/// or slower (more CPU burn) depending on the content.
- /// </item>
+ /// </description></item>
/// </list>
/// </summary>
/// <param name="tpv"></param>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Highlighter/PostingsHighlight/PassageScorer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Highlighter/PostingsHighlight/PassageScorer.cs b/src/Lucene.Net.Highlighter/PostingsHighlight/PassageScorer.cs
index de0fd45..fc64a70 100644
--- a/src/Lucene.Net.Highlighter/PostingsHighlight/PassageScorer.cs
+++ b/src/Lucene.Net.Highlighter/PostingsHighlight/PassageScorer.cs
@@ -43,9 +43,9 @@ namespace Lucene.Net.Search.PostingsHighlight
/// <summary>
/// Creates <see cref="PassageScorer"/> with these default values:
/// <list type="bullet">
- /// <item><c>k1 = 1.2</c></item>
- /// <item><c>b = 0.75</c></item>
- /// <item><c>pivot = 87</c></item>
+ /// <item><description><c>k1 = 1.2</c></description></item>
+ /// <item><description><c>b = 0.75</c></description></item>
+ /// <item><description><c>pivot = 87</c></description></item>
/// </list>
/// </summary>
public PassageScorer()
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Highlighter/PostingsHighlight/PostingsHighlighter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Highlighter/PostingsHighlight/PostingsHighlighter.cs b/src/Lucene.Net.Highlighter/PostingsHighlight/PostingsHighlighter.cs
index 6211042..7562228 100644
--- a/src/Lucene.Net.Highlighter/PostingsHighlight/PostingsHighlighter.cs
+++ b/src/Lucene.Net.Highlighter/PostingsHighlight/PostingsHighlighter.cs
@@ -46,10 +46,10 @@ namespace Lucene.Net.Search.PostingsHighlight
/// <para/>
/// You can customize the behavior by subclassing this highlighter, some important hooks:
/// <list type="bullet">
- /// <item><see cref="GetBreakIterator(string)"/>: Customize how the text is divided into passages.</item>
- /// <item><see cref="GetScorer(string)"/>: Customize how passages are ranked.</item>
- /// <item><see cref="GetFormatter(string)"/>: Customize how snippets are formatted.</item>
- /// <item><see cref="GetIndexAnalyzer(string)"/>: Enable highlighting of MultiTermQuerys such as <see cref="WildcardQuery"/>.</item>
+ /// <item><see cref="GetBreakIterator(string)"/>: Customize how the text is divided into passages.</description></item>
+ /// <item><see cref="GetScorer(string)"/>: Customize how passages are ranked.</description></item>
+ /// <item><see cref="GetFormatter(string)"/>: Customize how snippets are formatted.</description></item>
+ /// <item><see cref="GetIndexAnalyzer(string)"/>: Enable highlighting of MultiTermQuerys such as <see cref="WildcardQuery"/>.</description></item>
/// </list>
/// <para/>
/// <b>WARNING</b>: The code is very new and probably still has some exciting bugs!
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Misc/Index/Sorter/BlockJoinComparatorSource.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Misc/Index/Sorter/BlockJoinComparatorSource.cs b/src/Lucene.Net.Misc/Index/Sorter/BlockJoinComparatorSource.cs
index 1ee6f1c..db9b8d1 100644
--- a/src/Lucene.Net.Misc/Index/Sorter/BlockJoinComparatorSource.cs
+++ b/src/Lucene.Net.Misc/Index/Sorter/BlockJoinComparatorSource.cs
@@ -28,9 +28,9 @@ namespace Lucene.Net.Index.Sorter
/// Note that this class is intended to used with <see cref="SortingMergePolicy"/>,
/// and for other purposes has some limitations:
/// <list type="bullet">
- /// <item>Cannot yet be used with <see cref="IndexSearcher.SearchAfter(ScoreDoc, Query, Filter, int, Sort)">
- /// IndexSearcher.SearchAfter</see></item>
- /// <item>Filling sort field values is not yet supported.</item>
+ /// <item><description>Cannot yet be used with <see cref="IndexSearcher.SearchAfter(ScoreDoc, Query, Filter, int, Sort)">
+ /// IndexSearcher.SearchAfter</see></description></item>
+ /// <item><description>Filling sort field values is not yet supported.</description></item>
/// </list>
/// @lucene.experimental
/// </para>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Queries/CustomScoreQuery.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Queries/CustomScoreQuery.cs b/src/Lucene.Net.Queries/CustomScoreQuery.cs
index 1ee6639..e997f3c 100644
--- a/src/Lucene.Net.Queries/CustomScoreQuery.cs
+++ b/src/Lucene.Net.Queries/CustomScoreQuery.cs
@@ -29,8 +29,8 @@ namespace Lucene.Net.Queries
/// <summary>
/// Query that sets document score as a programmatic function of several (sub) scores:
/// <list type="bullet">
- /// <item>the score of its subQuery (any query)</item>
- /// <item>(optional) the score of its <see cref="FunctionQuery"/> (or queries).</item>
+ /// <item>the score of its subQuery (any query)</description></item>
+ /// <item>(optional) the score of its <see cref="FunctionQuery"/> (or queries).</description></item>
/// </list>
/// Subclasses can modify the computation by overriding <see cref="GetCustomScoreProvider"/>.
///
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Queries/Mlt/MoreLikeThis.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Queries/Mlt/MoreLikeThis.cs b/src/Lucene.Net.Queries/Mlt/MoreLikeThis.cs
index 16b2fe0..f9d0312 100644
--- a/src/Lucene.Net.Queries/Mlt/MoreLikeThis.cs
+++ b/src/Lucene.Net.Queries/Mlt/MoreLikeThis.cs
@@ -86,11 +86,11 @@ namespace Lucene.Net.Queries.Mlt
/// <para/>
/// Thus you:
/// <list type="bullet">
- /// <item>do your normal, Lucene setup for searching,</item>
- /// <item>create a MoreLikeThis,</item>
- /// <item>get the text of the doc you want to find similarities to</item>
- /// <item>then call one of the <see cref="Like"/> calls to generate a similarity query</item>
- /// <item>call the searcher to find the similar docs</item>
+ /// <item><description>do your normal, Lucene setup for searching,</description></item>
+ /// <item><description>create a MoreLikeThis,</description></item>
+ /// <item><description>get the text of the doc you want to find similarities to</description></item>
+ /// <item><description>then call one of the <see cref="Like"/> calls to generate a similarity query</description></item>
+ /// <item><description>call the searcher to find the similar docs</description></item>
/// </list>
/// <para/>
/// <b>More Advanced Usage</b>
@@ -103,15 +103,15 @@ namespace Lucene.Net.Queries.Mlt
/// may want to call the other set methods to control how the similarity queries are
/// generated:
/// <list type="bullet">
- /// <item><see cref="MinTermFreq"/></item>
- /// <item><see cref="MinDocFreq"/></item>
- /// <item><see cref="MaxDocFreq"/></item>
- /// <item><see cref="SetMaxDocFreqPct(int)"/></item>
- /// <item><see cref="MinWordLen"/></item>
- /// <item><see cref="MaxWordLen"/></item>
- /// <item><see cref="MaxQueryTerms"/></item>
- /// <item><see cref="MaxNumTokensParsed"/></item>
- /// <item><see cref="StopWords"/></item>
+ /// <item><description><see cref="MinTermFreq"/></description></item>
+ /// <item><description><see cref="MinDocFreq"/></description></item>
+ /// <item><description><see cref="MaxDocFreq"/></description></item>
+ /// <item><description><see cref="SetMaxDocFreqPct(int)"/></description></item>
+ /// <item><description><see cref="MinWordLen"/></description></item>
+ /// <item><description><see cref="MaxWordLen"/></description></item>
+ /// <item><description><see cref="MaxQueryTerms"/></description></item>
+ /// <item><description><see cref="MaxNumTokensParsed"/></description></item>
+ /// <item><description><see cref="StopWords"/></description></item>
/// </list>
/// </summary>
/// <remarks>
@@ -650,12 +650,12 @@ namespace Lucene.Net.Queries.Mlt
/// Each array has 6 elements.
/// The elements are:
/// <list type="bullet">
- /// <item>The word (<see cref="string"/>)</item>
- /// <item>The top field that this word comes from (<see cref="string"/>)</item>
- /// <item>The score for this word (<see cref="float"/>)</item>
- /// <item>The IDF value (<see cref="float"/>)</item>
- /// <item>The frequency of this word in the index (<see cref="int"/>)</item>
- /// <item>The frequency of this word in the source document (<see cref="int"/>)</item>
+ /// <item><description>The word (<see cref="string"/>)</description></item>
+ /// <item><description>The top field that this word comes from (<see cref="string"/>)</description></item>
+ /// <item><description>The score for this word (<see cref="float"/>)</description></item>
+ /// <item><description>The IDF value (<see cref="float"/>)</description></item>
+ /// <item><description>The frequency of this word in the index (<see cref="int"/>)</description></item>
+ /// <item><description>The frequency of this word in the source document (<see cref="int"/>)</description></item>
/// </list>
/// This is a somewhat "advanced" routine, and in general only the 1st entry in the array is of interest.
/// This method is exposed so that you can identify the "interesting words" in a document.
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.QueryParser/Classic/QueryParser.cs b/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
index 070802a..3a35db8 100644
--- a/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
+++ b/src/Lucene.Net.QueryParser/Classic/QueryParser.cs
@@ -34,19 +34,19 @@ namespace Lucene.Net.QueryParsers.Classic
/// A Query is a series of clauses.
/// A clause may be prefixed by:
/// <list type="bullet">
- /// <item> a plus (<c>+</c>) or a minus (<c>-</c>) sign, indicating
- /// that the clause is required or prohibited respectively; or</item>
- /// <item> a term followed by a colon, indicating the field to be searched.
- /// This enables one to construct queries which search multiple fields.</item>
+ /// <item><description> a plus (<c>+</c>) or a minus (<c>-</c>) sign, indicating
+ /// that the clause is required or prohibited respectively; or</description></item>
+ /// <item><description> a term followed by a colon, indicating the field to be searched.
+ /// This enables one to construct queries which search multiple fields.</description></item>
/// </list>
///
/// <para/>
/// A clause may be either:
/// <list type="bullet">
- /// <item> a term, indicating all the documents that contain this term; or</item>
- /// <item> a nested query, enclosed in parentheses. Note that this may be used
+ /// <item><description> a term, indicating all the documents that contain this term; or</description></item>
+ /// <item><description> a nested query, enclosed in parentheses. Note that this may be used
/// with a <c>+</c>/<c>-</c> prefix to require any of a set of
- /// terms.</item>
+ /// terms.</description></item>
/// </list>
///
/// <para/>
@@ -95,7 +95,7 @@ namespace Lucene.Net.QueryParsers.Classic
/// <b>NOTE</b>: You must specify the required <see cref="LuceneVersion" /> compatibility when
/// creating QueryParser:
/// <list type="bullet">
- /// <item>As of 3.1, <see cref="QueryParserBase.AutoGeneratePhraseQueries"/> is false by default.</item>
+ /// <item><description>As of 3.1, <see cref="QueryParserBase.AutoGeneratePhraseQueries"/> is false by default.</description></item>
/// </list>
/// </summary>
public class QueryParser : QueryParserBase
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.QueryParser/Flexible/Core/QueryParserHelper.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.QueryParser/Flexible/Core/QueryParserHelper.cs b/src/Lucene.Net.QueryParser/Flexible/Core/QueryParserHelper.cs
index 8bc7be4..f113c99 100644
--- a/src/Lucene.Net.QueryParser/Flexible/Core/QueryParserHelper.cs
+++ b/src/Lucene.Net.QueryParser/Flexible/Core/QueryParserHelper.cs
@@ -195,19 +195,19 @@ namespace Lucene.Net.QueryParsers.Flexible.Core
/// In this method the three phases are executed:
/// <para/>
/// <list type="number">
- /// <item>
+ /// <item><description>
/// the query string is parsed using the
/// text parser returned by <see cref="SyntaxParser"/>, the result is a query
/// node tree.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// the query node tree is processed by the
/// processor returned by <see cref="QueryNodeProcessor"/>.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// a object is built from the query node
/// tree using the builder returned by <see cref="QueryBuilder"/>.
- /// </item>
+ /// </description></item>
/// </list>
/// </summary>
/// <param name="query">the query string</param>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.QueryParser/Flexible/Standard/StandardQueryParser.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.QueryParser/Flexible/Standard/StandardQueryParser.cs b/src/Lucene.Net.QueryParser/Flexible/Standard/StandardQueryParser.cs
index 1940263..dedcd9f 100644
--- a/src/Lucene.Net.QueryParser/Flexible/Standard/StandardQueryParser.cs
+++ b/src/Lucene.Net.QueryParser/Flexible/Standard/StandardQueryParser.cs
@@ -54,25 +54,25 @@ namespace Lucene.Net.QueryParsers.Flexible.Standard
/// <para/>
/// A Query is a series of clauses. A clause may be prefixed by:
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// a plus (<c>+</c>) or a minus (<c>-</c>) sign, indicating that
/// the clause is required or prohibited respectively; or
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// a term followed by a colon, indicating the field to be searched. This
/// enables one to construct queries which search multiple fields.
- /// </item>
+ /// </description></item>
/// </list>
///
/// A clause may be either:
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// a term, indicating all the documents that contain this term; or
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// a nested query, enclosed in parentheses. Note that this may be used with
/// a <c>+</c>/<c>-</c> prefix to require any of a set of terms.
- /// </item>
+ /// </description></item>
/// </list>
///
/// Thus, in BNF, the query grammar is:
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.QueryParser/Simple/SimpleQueryParser.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.QueryParser/Simple/SimpleQueryParser.cs b/src/Lucene.Net.QueryParser/Simple/SimpleQueryParser.cs
index 01af234..7438bdd 100644
--- a/src/Lucene.Net.QueryParser/Simple/SimpleQueryParser.cs
+++ b/src/Lucene.Net.QueryParser/Simple/SimpleQueryParser.cs
@@ -69,14 +69,14 @@ namespace Lucene.Net.QueryParsers.Simple
/// to decipher what it can; however, this may mean odd or unexpected results.
/// <h4>Query Operators</h4>
/// <list type="bullet">
- /// <item>'<c>+</c>' specifies <c>AND</c> operation: <c>token1+token2</c></item>
- /// <item>'<c>|</c>' specifies <c>OR</c> operation: <c>token1|token2</c></item>
- /// <item>'<c>-</c>' negates a single token: <c>-token0</c></item>
- /// <item>'<c>"</c>' creates phrases of terms: <c>"term1 term2 ..."</c></item>
- /// <item>'<c>*</c>' at the end of terms specifies prefix query: <c>term*</c></item>
- /// <item>'<c>~</c>N' at the end of terms specifies fuzzy query: <c>term~1</c></item>
- /// <item>'<c>~</c>N' at the end of phrases specifies near query: <c>"term1 term2"~5</c></item>
- /// <item>'<c>(</c>' and '<c>)</c>' specifies precedence: <c>token1 + (token2 | token3)</c></item>
+ /// <item><description>'<c>+</c>' specifies <c>AND</c> operation: <c>token1+token2</c></description></item>
+ /// <item><description>'<c>|</c>' specifies <c>OR</c> operation: <c>token1|token2</c></description></item>
+ /// <item><description>'<c>-</c>' negates a single token: <c>-token0</c></description></item>
+ /// <item><description>'<c>"</c>' creates phrases of terms: <c>"term1 term2 ..."</c></description></item>
+ /// <item><description>'<c>*</c>' at the end of terms specifies prefix query: <c>term*</c></description></item>
+ /// <item><description>'<c>~</c>N' at the end of terms specifies fuzzy query: <c>term~1</c></description></item>
+ /// <item><description>'<c>~</c>N' at the end of phrases specifies near query: <c>"term1 term2"~5</c></description></item>
+ /// <item><description>'<c>(</c>' and '<c>)</c>' specifies precedence: <c>token1 + (token2 | token3)</c></description></item>
/// </list>
/// <para/>
/// The default operator is <c>OR</c> if no other operator is specified.
@@ -99,20 +99,20 @@ namespace Lucene.Net.QueryParsers.Simple
/// beyond the first character do not need to be escaped.
/// For example:
/// <list type="bullet">
- /// <item><c>-term1</c> -- Specifies <c>NOT</c> operation against <c>term1</c></item>
- /// <item><c>\-term1</c> -- Searches for the term <c>-term1</c>.</item>
- /// <item><c>term-1</c> -- Searches for the term <c>term-1</c>.</item>
- /// <item><c>term\-1</c> -- Searches for the term <c>term-1</c>.</item>
+ /// <item><description><c>-term1</c> -- Specifies <c>NOT</c> operation against <c>term1</c></description></item>
+ /// <item><description><c>\-term1</c> -- Searches for the term <c>-term1</c>.</description></item>
+ /// <item><description><c>term-1</c> -- Searches for the term <c>term-1</c>.</description></item>
+ /// <item><description><c>term\-1</c> -- Searches for the term <c>term-1</c>.</description></item>
/// </list>
/// <para/>
/// The '<c>*</c>' operator is a special case. On individual terms (not phrases) the last
/// character of a term that is '<c>*</c>' must be escaped; however, any '<c>*</c>' characters
/// before the last character do not need to be escaped:
/// <list type="bullet">
- /// <item><c>term1*</c> -- Searches for the prefix <c>term1</c></item>
- /// <item><c>term1\*</c> -- Searches for the term <c>term1*</c></item>
- /// <item><c>term*1</c> -- Searches for the term <c>term*1</c></item>
- /// <item><c>term\*1</c> -- Searches for the term <c>term*1</c></item>
+ /// <item><description><c>term1*</c> -- Searches for the prefix <c>term1</c></description></item>
+ /// <item><description><c>term1\*</c> -- Searches for the term <c>term1*</c></description></item>
+ /// <item><description><c>term*1</c> -- Searches for the term <c>term*1</c></description></item>
+ /// <item><description><c>term\*1</c> -- Searches for the term <c>term*1</c></description></item>
/// </list>
/// <para/>
/// Note that above examples consider the terms before text processing.
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.QueryParser/Surround/Query/SpanNearClauseFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.QueryParser/Surround/Query/SpanNearClauseFactory.cs b/src/Lucene.Net.QueryParser/Surround/Query/SpanNearClauseFactory.cs
index 7fba250..8fbc219 100644
--- a/src/Lucene.Net.QueryParser/Surround/Query/SpanNearClauseFactory.cs
+++ b/src/Lucene.Net.QueryParser/Surround/Query/SpanNearClauseFactory.cs
@@ -29,13 +29,13 @@ namespace Lucene.Net.QueryParsers.Surround.Query
/// Operations:
///
/// <list type="bullet">
- /// <item>create for a field name and an indexreader.</item>
+ /// <item><description>create for a field name and an indexreader.</description></item>
///
- /// <item>add a weighted Term - this should add a corresponding SpanTermQuery, or increase the weight of an existing one.</item>
+ /// <item><description>add a weighted Term - this should add a corresponding SpanTermQuery, or increase the weight of an existing one.</description></item>
///
- /// <item>add a weighted subquery SpanNearQuery</item>
+ /// <item><description>add a weighted subquery SpanNearQuery</description></item>
///
- /// <item>create a clause for SpanNearQuery from the things added above.</item>
+ /// <item><description>create a clause for SpanNearQuery from the things added above.</description></item>
/// </list>
/// <para/>
/// For this, create an array of SpanQuery's from the added ones.
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Sandbox/Queries/SortedSetSortField.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Sandbox/Queries/SortedSetSortField.cs b/src/Lucene.Net.Sandbox/Queries/SortedSetSortField.cs
index 78a27db..2069099 100644
--- a/src/Lucene.Net.Sandbox/Queries/SortedSetSortField.cs
+++ b/src/Lucene.Net.Sandbox/Queries/SortedSetSortField.cs
@@ -39,15 +39,15 @@ namespace Lucene.Net.Sandbox.Queries
/// <para/>
/// Limitations:
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// Fields containing <see cref="int.MaxValue"/> or more unique values
/// are unsupported.
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// Selectors other than the default <see cref="Selector.MIN"/> require
/// optional codec support. However several codecs provided by Lucene,
/// including the current default codec, support this.
- /// </item>
+ /// </description></item>
/// </list>
/// </summary>
public class SortedSetSortField : SortField
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Spatial/Prefix/PrefixTreeStrategy.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Spatial/Prefix/PrefixTreeStrategy.cs b/src/Lucene.Net.Spatial/Prefix/PrefixTreeStrategy.cs
index 61ed934..ac3d79e 100644
--- a/src/Lucene.Net.Spatial/Prefix/PrefixTreeStrategy.cs
+++ b/src/Lucene.Net.Spatial/Prefix/PrefixTreeStrategy.cs
@@ -41,27 +41,27 @@ namespace Lucene.Net.Spatial.Prefix
///
/// <h4>Characteristics:</h4>
/// <list type="bullet">
- /// <item>Can index any shape; however only
+ /// <item><description>Can index any shape; however only
/// <see cref="RecursivePrefixTreeStrategy">RecursivePrefixTreeStrategy</see>
- /// can effectively search non-point shapes.</item>
- /// <item>Can index a variable number of shapes per field value. This strategy
+ /// can effectively search non-point shapes.</description></item>
+ /// <item><description>Can index a variable number of shapes per field value. This strategy
/// can do it via multiple calls to <see cref="CreateIndexableFields(IShape)"/>
/// for a document or by giving it some sort of Shape aggregate (e.g. NTS
/// WKT MultiPoint). The shape's boundary is approximated to a grid precision.
- /// </item>
- /// <item>Can query with any shape. The shape's boundary is approximated to a grid
- /// precision.</item>
- /// <item>Only <see cref="SpatialOperation.Intersects"/>
+ /// </description></item>
+ /// <item><description>Can query with any shape. The shape's boundary is approximated to a grid
+ /// precision.</description></item>
+ /// <item><description>Only <see cref="SpatialOperation.Intersects"/>
/// is supported. If only points are indexed then this is effectively equivalent
- /// to IsWithin.</item>
- /// <item>The strategy supports <see cref="MakeDistanceValueSource(IPoint, double)"/>
+ /// to IsWithin.</description></item>
+ /// <item><description>The strategy supports <see cref="MakeDistanceValueSource(IPoint, double)"/>
/// even for multi-valued data, so long as the indexed data is all points; the
/// behavior is undefined otherwise. However, <c>it will likely be removed in
/// the future</c> in lieu of using another strategy with a more scalable
/// implementation. Use of this call is the only
/// circumstance in which a cache is used. The cache is simple but as such
/// it doesn't scale to large numbers of points nor is it real-time-search
- /// friendly.</item>
+ /// friendly.</description></item>
/// </list>
///
/// <h4>Implementation:</h4>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Spatial/SpatialStrategy.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Spatial/SpatialStrategy.cs b/src/Lucene.Net.Spatial/SpatialStrategy.cs
index 348ca7e..bb39500 100644
--- a/src/Lucene.Net.Spatial/SpatialStrategy.cs
+++ b/src/Lucene.Net.Spatial/SpatialStrategy.cs
@@ -32,11 +32,11 @@ namespace Lucene.Net.Spatial
/// Different implementations will support different features. A strategy should
/// document these common elements:
/// <list type="bullet">
- /// <item>Can it index more than one shape per field?</item>
- /// <item>What types of shapes can be indexed?</item>
- /// <item>What types of query shapes can be used?</item>
- /// <item>What types of query operations are supported? This might vary per shape.</item>
- /// <item>Does it use the <see cref="FieldCache"/>, or some other type of cache? When?</item>
+ /// <item><description>Can it index more than one shape per field?</description></item>
+ /// <item><description>What types of shapes can be indexed?</description></item>
+ /// <item><description>What types of query shapes can be used?</description></item>
+ /// <item><description>What types of query operations are supported? This might vary per shape.</description></item>
+ /// <item><description>Does it use the <see cref="FieldCache"/>, or some other type of cache? When?</description></item>
/// </list>
/// If a strategy only supports certain shapes at index or query time, then in
/// general it will throw an exception if given an incompatible one. It will not
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Spatial/Vector/PointVectorStrategy.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Spatial/Vector/PointVectorStrategy.cs b/src/Lucene.Net.Spatial/Vector/PointVectorStrategy.cs
index e8a0e6a..28a0050 100644
--- a/src/Lucene.Net.Spatial/Vector/PointVectorStrategy.cs
+++ b/src/Lucene.Net.Spatial/Vector/PointVectorStrategy.cs
@@ -32,11 +32,11 @@ namespace Lucene.Net.Spatial.Vector
///
/// <h4>Characteristics:</h4>
/// <list type="bullet">
- /// <item>Only indexes points; just one per field value.</item>
- /// <item>Can query by a rectangle or circle.</item>
- /// <item><see cref="SpatialOperation.Intersects"/> and <see cref="SpatialOperation.IsWithin"/> is supported.</item>
- /// <item>Uses the FieldCache for <see cref="SpatialStrategy.MakeDistanceValueSource(IPoint)"/> and for
- /// searching with a Circle.</item>
+ /// <item><description>Only indexes points; just one per field value.</description></item>
+ /// <item><description>Can query by a rectangle or circle.</description></item>
+ /// <item><description><see cref="SpatialOperation.Intersects"/> and <see cref="SpatialOperation.IsWithin"/> is supported.</description></item>
+ /// <item><description>Uses the FieldCache for <see cref="SpatialStrategy.MakeDistanceValueSource(IPoint)"/> and for
+ /// searching with a Circle.</description></item>
/// </list>
///
/// <h4>Implementation:</h4>
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Suggest/Spell/LuceneLevenshteinDistance.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Suggest/Spell/LuceneLevenshteinDistance.cs b/src/Lucene.Net.Suggest/Spell/LuceneLevenshteinDistance.cs
index 5f8988b..4010f80 100644
--- a/src/Lucene.Net.Suggest/Spell/LuceneLevenshteinDistance.cs
+++ b/src/Lucene.Net.Suggest/Spell/LuceneLevenshteinDistance.cs
@@ -27,13 +27,13 @@ namespace Lucene.Net.Search.Spell
///
/// Notes:
/// <list type="bullet">
- /// <item> This metric treats full unicode codepoints as characters</item>
- /// <item> This metric scales raw edit distances into a floating point score
- /// based upon the shortest of the two terms</item>
- /// <item> Transpositions of two adjacent codepoints are treated as primitive
- /// edits.</item>
- /// <item> Edits are applied in parallel: for example, "ab" and "bca" have
- /// distance 3.</item>
+ /// <item><description> This metric treats full unicode codepoints as characters</description></item>
+ /// <item><description> This metric scales raw edit distances into a floating point score
+ /// based upon the shortest of the two terms</description></item>
+ /// <item><description> Transpositions of two adjacent codepoints are treated as primitive
+ /// edits.</description></item>
+ /// <item><description> Edits are applied in parallel: for example, "ab" and "bca" have
+ /// distance 3.</description></item>
/// </list>
///
/// NOTE: this class is not particularly efficient. It is only intended
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Suggest/Suggest/Analyzing/AnalyzingSuggester.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Suggest/Suggest/Analyzing/AnalyzingSuggester.cs b/src/Lucene.Net.Suggest/Suggest/Analyzing/AnalyzingSuggester.cs
index 8c58e5a..0f85629 100644
--- a/src/Lucene.Net.Suggest/Suggest/Analyzing/AnalyzingSuggester.cs
+++ b/src/Lucene.Net.Suggest/Suggest/Analyzing/AnalyzingSuggester.cs
@@ -65,21 +65,21 @@ namespace Lucene.Net.Search.Suggest.Analyzing
/// There are some limitations:
/// <list type="number">
///
- /// <item> A lookup from a query like "net" in English won't
+ /// <item><description> A lookup from a query like "net" in English won't
/// be any different than "net " (ie, user added a
/// trailing space) because analyzers don't reflect
/// when they've seen a token separator and when they
- /// haven't.</item>
+ /// haven't.</description></item>
///
- /// <item> If you're using <see cref="Analysis.Core.StopFilter"/>, and the user will
+ /// <item><description> If you're using <see cref="Analysis.Core.StopFilter"/>, and the user will
/// type "fast apple", but so far all they've typed is
/// "fast a", again because the analyzer doesn't convey whether
/// it's seen a token separator after the "a",
/// <see cref="Analysis.Core.StopFilter"/> will remove that "a" causing
- /// far more matches than you'd expect.</item>
+ /// far more matches than you'd expect.</description></item>
///
- /// <item> Lookups with the empty string return no results
- /// instead of all results.</item>
+ /// <item><description> Lookups with the empty string return no results
+ /// instead of all results.</description></item>
/// </list>
///
/// @lucene.experimental
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Suggest/Suggest/DocumentDictionary.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Suggest/Suggest/DocumentDictionary.cs b/src/Lucene.Net.Suggest/Suggest/DocumentDictionary.cs
index 4026e36..af9fabe 100644
--- a/src/Lucene.Net.Suggest/Suggest/DocumentDictionary.cs
+++ b/src/Lucene.Net.Suggest/Suggest/DocumentDictionary.cs
@@ -31,19 +31,19 @@ namespace Lucene.Net.Search.Suggest
/// </para>
/// <b>NOTE:</b>
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// The term and (optionally) payload fields have to be
/// stored
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// The weight field can be stored or can be a <see cref="NumericDocValues"/>.
/// If the weight field is not defined, the value of the weight is <c>0</c>
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// if any of the term or (optionally) payload fields supplied
/// do not have a value for a document, then the document is
/// skipped by the dictionary
- /// </item>
+ /// </description></item>
/// </list>
/// </summary>
public class DocumentDictionary : IDictionary
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Suggest/Suggest/DocumentValueSourceDictionary.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Suggest/Suggest/DocumentValueSourceDictionary.cs b/src/Lucene.Net.Suggest/Suggest/DocumentValueSourceDictionary.cs
index 401e46a..baf0e9f 100644
--- a/src/Lucene.Net.Suggest/Suggest/DocumentValueSourceDictionary.cs
+++ b/src/Lucene.Net.Suggest/Suggest/DocumentValueSourceDictionary.cs
@@ -33,15 +33,15 @@ namespace Lucene.Net.Search.Suggest
/// </para>
/// <b>NOTE:</b>
/// <list type="bullet">
- /// <item>
+ /// <item><description>
/// The term and (optionally) payload fields have to be
/// stored
- /// </item>
- /// <item>
+ /// </description></item>
+ /// <item><description>
/// if the term or (optionally) payload fields supplied
/// do not have a value for a document, then the document is
/// rejected by the dictionary
- /// </item>
+ /// </description></item>
/// </list>
/// <para>
/// In practice the <see cref="ValueSource"/> will likely be obtained
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Suggest/Suggest/FileDictionary.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Suggest/Suggest/FileDictionary.cs b/src/Lucene.Net.Suggest/Suggest/FileDictionary.cs
index e6f48da..35d6ab0 100644
--- a/src/Lucene.Net.Suggest/Suggest/FileDictionary.cs
+++ b/src/Lucene.Net.Suggest/Suggest/FileDictionary.cs
@@ -30,20 +30,20 @@ namespace Lucene.Net.Search.Suggest
/// <para>Format allowed: 1 entry per line:</para>
/// <para>An entry can be: </para>
/// <list type="number">
- /// <item>suggestion</item>
- /// <item>suggestion <see cref="fieldDelimiter"/> weight</item>
- /// <item>suggestion <see cref="fieldDelimiter"/> weight <see cref="fieldDelimiter"/> payload</item>
+ /// <item><description>suggestion</description></item>
+ /// <item><description>suggestion <see cref="fieldDelimiter"/> weight</description></item>
+ /// <item><description>suggestion <see cref="fieldDelimiter"/> weight <see cref="fieldDelimiter"/> payload</description></item>
/// </list>
/// where the default <see cref="fieldDelimiter"/> is <see cref="DEFAULT_FIELD_DELIMITER"/> (a tab)
/// <para>
/// <b>NOTE:</b>
/// <list type="number">
- /// <item>In order to have payload enabled, the first entry has to have a payload</item>
- /// <item>If the weight for an entry is not specified then a value of 1 is used</item>
- /// <item>A payload cannot be specified without having the weight specified for an entry</item>
- /// <item>If the payload for an entry is not specified (assuming payload is enabled)
- /// then an empty payload is returned</item>
- /// <item>An entry cannot have more than two <see cref="fieldDelimiter"/>s</item>
+ /// <item><description>In order to have payload enabled, the first entry has to have a payload</description></item>
+ /// <item><description>If the weight for an entry is not specified then a value of 1 is used</description></item>
+ /// <item><description>A payload cannot be specified without having the weight specified for an entry</description></item>
+ /// <item><description>If the payload for an entry is not specified (assuming payload is enabled)
+ /// then an empty payload is returned</description></item>
+ /// <item><description>An entry cannot have more than two <see cref="fieldDelimiter"/>s</description></item>
/// </list>
/// </para>
/// <c>Example:</c><para/>