You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucenenet.apache.org by sy...@apache.org on 2015/01/28 12:11:50 UTC

[1/2] lucenenet git commit: Various

Repository: lucenenet
Updated Branches:
  refs/heads/master b5be9c0b8 -> 67af7e4e9


Various


Project: http://git-wip-us.apache.org/repos/asf/lucenenet/repo
Commit: http://git-wip-us.apache.org/repos/asf/lucenenet/commit/4513c521
Tree: http://git-wip-us.apache.org/repos/asf/lucenenet/tree/4513c521
Diff: http://git-wip-us.apache.org/repos/asf/lucenenet/diff/4513c521

Branch: refs/heads/master
Commit: 4513c521802861b3c3ebe4e33647a4670f035727
Parents: b5be9c0
Author: Itamar Syn-Hershko <it...@code972.com>
Authored: Wed Jan 28 13:10:59 2015 +0200
Committer: Itamar Syn-Hershko <it...@code972.com>
Committed: Wed Jan 28 13:11:33 2015 +0200

----------------------------------------------------------------------
 src/Lucene.Net.Core/Analysis/Analyzer.cs        | 28 +++-----
 .../Analysis/NumericTokenStream.cs              | 67 ++++++++------------
 .../Analysis/ReusableStringReader.cs            | 34 +++++-----
 src/Lucene.Net.Core/Analysis/Token.cs           | 31 +++++----
 src/Lucene.Net.Core/Analysis/TokenStream.cs     |  4 +-
 src/Lucene.Net.Core/Analysis/Tokenizer.cs       | 12 ++--
 src/Lucene.Net.Core/Support/WeakDictionary.cs   |  4 +-
 7 files changed, 79 insertions(+), 101 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Analysis/Analyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/Analyzer.cs b/src/Lucene.Net.Core/Analysis/Analyzer.cs
index 9c9ab97..ce6d98e 100644
--- a/src/Lucene.Net.Core/Analysis/Analyzer.cs
+++ b/src/Lucene.Net.Core/Analysis/Analyzer.cs
@@ -68,7 +68,7 @@ namespace Lucene.Net.Analysis
     /// </summary>
     public abstract class Analyzer : IDisposable
     {
-        private readonly ReuseStrategy ReuseStrategy_Renamed;
+        private readonly ReuseStrategy _reuseStrategy;
 
         // non final as it gets nulled if closed; pkg private for access by ReuseStrategy's final helper methods:
         internal IDisposableThreadLocal<object> StoredValue = new IDisposableThreadLocal<object>();
@@ -77,7 +77,7 @@ namespace Lucene.Net.Analysis
         /// Create a new Analyzer, reusing the same set of components per-thread
         /// across calls to <seealso cref="#tokenStream(String, Reader)"/>.
         /// </summary>
-        public Analyzer()
+        protected Analyzer()
             : this(GLOBAL_REUSE_STRATEGY)
         {
         }
@@ -90,9 +90,9 @@ namespace Lucene.Net.Analysis
         /// <a href="{@docRoot}/../analyzers-common/Lucene.Net.Analysis/miscellaneous/PerFieldAnalyzerWrapper.html">
         /// PerFieldAnalyerWrapper</a> instead.
         /// </summary>
-        public Analyzer(ReuseStrategy reuseStrategy)
+        protected Analyzer(ReuseStrategy reuseStrategy)
         {
-            this.ReuseStrategy_Renamed = reuseStrategy;
+            this._reuseStrategy = reuseStrategy;
         }
 
         /// <summary>
@@ -129,16 +129,16 @@ namespace Lucene.Net.Analysis
         /// <seealso cref= #tokenStream(String, Reader) </seealso>
         public TokenStream TokenStream(string fieldName, TextReader reader)
         {
-            TokenStreamComponents components = ReuseStrategy_Renamed.GetReusableComponents(this, fieldName);
+            TokenStreamComponents components = _reuseStrategy.GetReusableComponents(this, fieldName);
             TextReader r = InitReader(fieldName, reader);
             if (components == null)
             {
                 components = CreateComponents(fieldName, r);
-                ReuseStrategy_Renamed.SetReusableComponents(this, fieldName, components);
+                _reuseStrategy.SetReusableComponents(this, fieldName, components);
             }
             else
             {
-                components.Reader = r;//new TextReaderWrapper(r);
+                components.Reader = r;// LUCENENET TODO new TextReaderWrapper(r);
             }
             return components.TokenStream;
         }
@@ -196,7 +196,7 @@ namespace Lucene.Net.Analysis
         {
             get
             {
-                return ReuseStrategy_Renamed;
+                return _reuseStrategy;
             }
         }
 
@@ -308,12 +308,6 @@ namespace Lucene.Net.Analysis
         public abstract class ReuseStrategy
         {
             /// <summary>
-            /// Sole constructor. (For invocation by subclass constructors, typically implicit.) </summary>
-            public ReuseStrategy()
-            {
-            }
-
-            /// <summary>
             /// Gets the reusable TokenStreamComponents for the field with the given name.
             /// </summary>
             /// <param name="analyzer"> Analyzer from which to get the reused components. Use
@@ -379,10 +373,6 @@ namespace Lucene.Net.Analysis
         /// Sole constructor. (For invocation by subclass constructors, typically implicit.) </summary>
         /// @deprecated Don't create instances of this class, use <seealso cref="Analyzer#GLOBAL_REUSE_STRATEGY"/>
         {
-            public GlobalReuseStrategy()
-            {
-            }
-
             public override TokenStreamComponents GetReusableComponents(Analyzer analyzer, string fieldName)
             {
                 return (TokenStreamComponents)GetStoredValue(analyzer);
@@ -418,9 +408,9 @@ namespace Lucene.Net.Analysis
             public override TokenStreamComponents GetReusableComponents(Analyzer analyzer, string fieldName)
             {
                 var componentsPerField = (IDictionary<string, TokenStreamComponents>)GetStoredValue(analyzer);
-                TokenStreamComponents ret;
                 if (componentsPerField != null)
                 {
+                    TokenStreamComponents ret;
                     componentsPerField.TryGetValue(fieldName, out ret);
                     return ret;
                 }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Analysis/NumericTokenStream.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/NumericTokenStream.cs b/src/Lucene.Net.Core/Analysis/NumericTokenStream.cs
index 1b218cb..fcb1e6d 100644
--- a/src/Lucene.Net.Core/Analysis/NumericTokenStream.cs
+++ b/src/Lucene.Net.Core/Analysis/NumericTokenStream.cs
@@ -1,4 +1,7 @@
+using System.Diagnostics;
 using Lucene.Net.Analysis.Tokenattributes;
+using Lucene.Net.Documents;
+using Lucene.Net.Search;
 using Lucene.Net.Util;
 using System;
 
@@ -136,7 +139,7 @@ namespace Lucene.Net.Analysis
         // just a wrapper to prevent adding CTA
         private sealed class NumericAttributeFactory : AttributeSource.AttributeFactory
         {
-            internal readonly AttributeSource.AttributeFactory @delegate;
+            private readonly AttributeSource.AttributeFactory @delegate;
 
             internal NumericAttributeFactory(AttributeSource.AttributeFactory @delegate)
             {
@@ -161,77 +164,59 @@ namespace Lucene.Net.Analysis
         /// </summary>
         public sealed class NumericTermAttribute : Util.Attribute, INumericTermAttribute, ITermToBytesRefAttribute
         {
-            internal long Value = 0L;
-            internal int ValueSize_Renamed = 0, Shift_Renamed = 0, PrecisionStep = 0;
-            internal BytesRef Bytes = new BytesRef();
+            private long _value = 0L;
+            private int _precisionStep = 0;
+            private readonly BytesRef _bytes = new BytesRef();
 
-            /// <summary>
-            /// Creates, but does not yet initialize this attribute instance </summary>
-            /// <seealso> cref= #init(long, int, int, int) </seealso>
             public NumericTermAttribute()
             {
+                ValueSize = 0;
             }
 
             public BytesRef BytesRef
             {
                 get
                 {
-                    return Bytes;
+                    return _bytes;
                 }
             }
 
             public void FillBytesRef()
             {
-                //Debug.Assert(ValueSize_Renamed == 64 || ValueSize_Renamed == 32);
-                if (ValueSize_Renamed == 64)
+                Debug.Assert(ValueSize == 64 || ValueSize == 32);
+                if (ValueSize == 64)
                 {
-                    NumericUtils.LongToPrefixCoded(Value, Shift_Renamed, Bytes);
+                    NumericUtils.LongToPrefixCoded(_value, Shift, _bytes);
                 }
                 else
                 {
-                    NumericUtils.IntToPrefixCoded((int)Value, Shift_Renamed, Bytes);
+                    NumericUtils.IntToPrefixCoded((int)_value, Shift, _bytes);
                 }
             }
 
-            public int Shift
-            {
-                get
-                {
-                    return Shift_Renamed;
-                }
-                set
-                {
-                    this.Shift_Renamed = value;
-                }
-            }
+            public int Shift { get; set; }
 
             public int IncShift()
             {
-                return (Shift_Renamed += PrecisionStep);
+                return (Shift += _precisionStep);
             }
 
             public long RawValue
             {
                 get
                 {
-                    return Value & ~((1L << Shift_Renamed) - 1L);
+                    return _value & ~((1L << Shift) - 1L);
                 }
             }
 
-            public int ValueSize
-            {
-                get
-                {
-                    return ValueSize_Renamed;
-                }
-            }
+            public int ValueSize { get; private set; }
 
             public void Init(long value, int valueSize, int precisionStep, int shift)
             {
-                this.Value = value;
-                this.ValueSize_Renamed = valueSize;
-                this.PrecisionStep = precisionStep;
-                this.Shift_Renamed = shift;
+                this._value = value;
+                this.ValueSize = valueSize;
+                this._precisionStep = precisionStep;
+                this.Shift = shift;
             }
 
             public override void Clear()
@@ -243,16 +228,16 @@ namespace Lucene.Net.Analysis
             public override void ReflectWith(IAttributeReflector reflector)
             {
                 FillBytesRef();
-                reflector.Reflect(typeof(ITermToBytesRefAttribute), "bytes", BytesRef.DeepCopyOf(Bytes));
-                reflector.Reflect(typeof(INumericTermAttribute), "shift", Shift_Renamed);
+                reflector.Reflect(typeof(ITermToBytesRefAttribute), "bytes", BytesRef.DeepCopyOf(_bytes));
+                reflector.Reflect(typeof(INumericTermAttribute), "shift", _shift);
                 reflector.Reflect(typeof(INumericTermAttribute), "rawValue", RawValue);
-                reflector.Reflect(typeof(INumericTermAttribute), "valueSize", ValueSize_Renamed);
+                reflector.Reflect(typeof(INumericTermAttribute), "valueSize", ValueSize);
             }
 
             public override void CopyTo(Util.Attribute target)
             {
-                NumericTermAttribute a = (NumericTermAttribute)target;
-                a.Init(Value, ValueSize_Renamed, PrecisionStep, Shift_Renamed);
+                var a = (NumericTermAttribute)target;
+                a.Init(_value, ValueSize, _precisionStep, _shift);
             }
         }
 

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Analysis/ReusableStringReader.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/ReusableStringReader.cs b/src/Lucene.Net.Core/Analysis/ReusableStringReader.cs
index 71b910a..76bb651 100644
--- a/src/Lucene.Net.Core/Analysis/ReusableStringReader.cs
+++ b/src/Lucene.Net.Core/Analysis/ReusableStringReader.cs
@@ -23,56 +23,52 @@ namespace Lucene.Net.Analysis
     /// Internal class to enable reuse of the string reader by <seealso cref="Analyzer#tokenStream(String,String)"/> </summary>
     public sealed class ReusableStringReader : System.IO.TextReader
     {
-        private int Pos = 0, Size = 0;
-        private string s = null;
-
-        public ReusableStringReader()
-        {
-        }
+        private int _pos = 0, _size = 0;
+        private string _s = null;
 
         public string Value
         {
             set
             {
-                this.s = value;
-                this.Size = value.Length;
-                this.Pos = 0;
+                this._s = value;
+                this._size = value.Length;
+                this._pos = 0;
             }
         }
 
         public override int Read()
         {
-            if (Pos < Size)
+            if (_pos < _size)
             {
-                return s[Pos++];
+                return _s[_pos++];
             }
             else
             {
-                s = null;
+                _s = null;
                 return -1;
             }
         }
 
         public override int Read(char[] c, int off, int len)
         {
-            if (Pos < Size)
+            if (_pos < _size)
             {
-                len = Math.Min(len, Size - Pos);
-                s.CopyTo(Pos, c, off, Pos + len - Pos);
-                Pos += len;
+                len = Math.Min(len, _size - _pos);
+                _s.CopyTo(_pos, c, off, _pos + len - _pos);
+                _pos += len;
                 return len;
             }
             else
             {
-                s = null;
+                _s = null;
                 return -1;
             }
         }
 
         public override void Close()
         {
-            Pos = Size; // this prevents NPE when reading after close!
-            s = null;
+            _pos = _size; // this prevents NPE when reading after close!
+            _s = null;
         }
     }
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Analysis/Token.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/Token.cs b/src/Lucene.Net.Core/Analysis/Token.cs
index b26f423..2a234b2 100644
--- a/src/Lucene.Net.Core/Analysis/Token.cs
+++ b/src/Lucene.Net.Core/Analysis/Token.cs
@@ -1,8 +1,8 @@
+using Lucene.Net.Index;
+
 namespace Lucene.Net.Analysis
 {
     using Lucene.Net.Analysis.Tokenattributes;
-
-    // for javadoc
     using Attribute = Lucene.Net.Util.Attribute;
     using AttributeSource = Lucene.Net.Util.AttributeSource;
     using BytesRef = Lucene.Net.Util.BytesRef;
@@ -67,7 +67,7 @@ namespace Lucene.Net.Analysis
     ///  if you know that your text is shorter than the capacity of the termBuffer
     ///  or <seealso cref="#resizeBuffer(int)"/>, if there is any possibility
     ///  that you may need to grow the buffer. Fill in the characters of your term into this
-    ///  buffer, with <seealso cref="String#getChars(int, int, char[], int)"/> if loading from a string,
+    ///  buffer, with <seealso cref="string#getChars(int, int, char[], int)"/> if loading from a string,
     ///  or with <seealso cref="System#arraycopy(Object, int, Object, int, int)"/>, and finally call <seealso cref="#setLength(int)"/> to
     ///  set the length of the term text.  See <a target="_top"
     ///  href="https://issues.apache.org/jira/browse/LUCENE-969">LUCENE-969</a>
@@ -353,7 +353,7 @@ namespace Lucene.Net.Analysis
 
         public override object Clone()
         {
-            Token t = (Token)base.Clone();
+            var t = (Token)base.Clone();
             // Do a deep clone
             if (payload != null)
             {
@@ -371,10 +371,12 @@ namespace Lucene.Net.Analysis
         /// </summary>
         public virtual Token Clone(char[] newTermBuffer, int newTermOffset, int newTermLength, int newStartOffset, int newEndOffset)
         {
-            Token t = new Token(newTermBuffer, newTermOffset, newTermLength, newStartOffset, newEndOffset);
-            t.positionIncrement = positionIncrement;
-            t.flags = flags;
-            t.type = type;
+            var t = new Token(newTermBuffer, newTermOffset, newTermLength, newStartOffset, newEndOffset)
+            {
+                positionIncrement = positionIncrement,
+                flags = flags,
+                type = type
+            };
             if (payload != null)
             {
                 t.payload = (BytesRef)payload.Clone();
@@ -389,9 +391,9 @@ namespace Lucene.Net.Analysis
                 return true;
             }
 
-            if (obj is Token)
+            var other = obj as Token;
+            if (other != null)
             {
-                Token other = (Token)obj;
                 return (startOffset == other.startOffset && endOffset == other.endOffset && flags == other.flags && positionIncrement == other.positionIncrement && (type == null ? other.type == null : type.Equals(other.type)) && (payload == null ? other.payload == null : payload.Equals(other.payload)) && base.Equals(obj));
             }
             else
@@ -580,9 +582,9 @@ namespace Lucene.Net.Analysis
 
         public override void CopyTo(Attribute target)
         {
-            if (target is Token)
+            var to = target as Token;
+            if (to != null)
             {
-                Token to = (Token)target;
                 to.Reinit(this);
                 // reinit shares the payload, so clone it:
                 if (payload != null)
@@ -658,9 +660,10 @@ namespace Lucene.Net.Analysis
                 {
                     return true;
                 }
-                if (other is TokenAttributeFactory)
+
+                var af = other as TokenAttributeFactory;
+                if (af != null)
                 {
-                    TokenAttributeFactory af = (TokenAttributeFactory)other;
                     return this.@delegate.Equals(af.@delegate);
                 }
                 return false;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Analysis/TokenStream.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/TokenStream.cs b/src/Lucene.Net.Core/Analysis/TokenStream.cs
index ccaed6b..c1c9d98 100644
--- a/src/Lucene.Net.Core/Analysis/TokenStream.cs
+++ b/src/Lucene.Net.Core/Analysis/TokenStream.cs
@@ -1,6 +1,8 @@
+using System.Diagnostics;
 using Lucene.Net.Analysis.Tokenattributes;
 using System;
 using Lucene.Net.Documents;
+using Lucene.Net.Index;
 using Lucene.Net.Util;
 
 namespace Lucene.Net.Analysis
@@ -107,7 +109,7 @@ namespace Lucene.Net.Analysis
             //Debug.Assert(AssertFinal());
         }
 
-        /* LUCENE TO-DO
+        /* LUCENENET TO-DO
         private bool AssertFinal()
         {
           try

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Analysis/Tokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/Tokenizer.cs b/src/Lucene.Net.Core/Analysis/Tokenizer.cs
index b63da43..b0215be 100644
--- a/src/Lucene.Net.Core/Analysis/Tokenizer.cs
+++ b/src/Lucene.Net.Core/Analysis/Tokenizer.cs
@@ -47,7 +47,7 @@ namespace Lucene.Net.Analysis
         {
             if (input == null)
             {
-                throw new System.NullReferenceException("input must not be null");
+                throw new System.ArgumentNullException("input", "input must not be null");
             }
             this.InputPending = input;
         }
@@ -59,7 +59,7 @@ namespace Lucene.Net.Analysis
         {
             if (input == null)
             {
-                throw new System.NullReferenceException("input must not be null");
+                throw new System.ArgumentNullException("input", "input must not be null");
             }
             this.InputPending = input;
         }
@@ -102,11 +102,11 @@ namespace Lucene.Net.Analysis
             {
                 if (value == null)
                 {
-                    throw new System.NullReferenceException("input must not be null");
+                    throw new System.ArgumentNullException("value", "input must not be null");
                 }
                 else if (this.input != ILLEGAL_STATE_READER)
                 {
-                    //throw new Exception("TokenStream contract violation: close() call missing");
+                    throw new InvalidOperationException("TokenStream contract violation: close() call missing");
                 }
                 this.InputPending = value;
                 Debug.Assert(SetReaderTestPoint());
@@ -132,7 +132,9 @@ namespace Lucene.Net.Analysis
         {
             public override int Read(char[] cbuf, int off, int len)
             {
-                throw new InvalidOperationException("TokenStream contract violation: reset()/close() call missing, " + "reset() called multiple times, or subclass does not call super.reset(). " + "Please see Javadocs of TokenStream class for more information about the correct consuming workflow.");
+                throw new InvalidOperationException("TokenStream contract violation: reset()/close() call missing, " 
+                    + "reset() called multiple times, or subclass does not call super.reset(). "
+                    + "Please see Javadocs of TokenStream class for more information about the correct consuming workflow.");
             }
 
             public override void Close()

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/4513c521/src/Lucene.Net.Core/Support/WeakDictionary.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Support/WeakDictionary.cs b/src/Lucene.Net.Core/Support/WeakDictionary.cs
index a29a0bf..c8b07de 100644
--- a/src/Lucene.Net.Core/Support/WeakDictionary.cs
+++ b/src/Lucene.Net.Core/Support/WeakDictionary.cs
@@ -26,7 +26,7 @@ using System.Linq;
 
 namespace Lucene.Net.Support
 {
-    public sealed class WeakDictionary<TKey, TValue> : IDictionary<TKey, TValue>
+    public sealed class WeakDictionary<TKey, TValue> : IDictionary<TKey, TValue> where TKey : class 
     {
         private HashMap<WeakKey<TKey>, TValue> _hm;
         private int _gcCollections = 0;
@@ -249,7 +249,7 @@ namespace Lucene.Net.Support
         /// is added to the hashtable, the key is wrapped using a WeakKey. WeakKey saves the
         /// value of the original object hashcode for fast comparison.
         /// </summary>
-        private class WeakKey<T>
+        private class WeakKey<T> where T : class
         {
             private readonly WeakReference reference;
             private readonly int hashCode;


[2/2] lucenenet git commit: We don't need a lock here

Posted by sy...@apache.org.
We don't need a lock here


Project: http://git-wip-us.apache.org/repos/asf/lucenenet/repo
Commit: http://git-wip-us.apache.org/repos/asf/lucenenet/commit/67af7e4e
Tree: http://git-wip-us.apache.org/repos/asf/lucenenet/tree/67af7e4e
Diff: http://git-wip-us.apache.org/repos/asf/lucenenet/diff/67af7e4e

Branch: refs/heads/master
Commit: 67af7e4e92937a7bb0aa13aab7599902cc14a5a3
Parents: 4513c52
Author: Itamar Syn-Hershko <it...@code972.com>
Authored: Wed Jan 28 13:11:09 2015 +0200
Committer: Itamar Syn-Hershko <it...@code972.com>
Committed: Wed Jan 28 13:11:34 2015 +0200

----------------------------------------------------------------------
 src/Lucene.Net.Core/Search/CachingWrapperFilter.cs | 10 +++-------
 1 file changed, 3 insertions(+), 7 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/lucenenet/blob/67af7e4e/src/Lucene.Net.Core/Search/CachingWrapperFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Search/CachingWrapperFilter.cs b/src/Lucene.Net.Core/Search/CachingWrapperFilter.cs
index 242225a..1399499 100644
--- a/src/Lucene.Net.Core/Search/CachingWrapperFilter.cs
+++ b/src/Lucene.Net.Core/Search/CachingWrapperFilter.cs
@@ -180,13 +180,9 @@ namespace Lucene.Net.Search
         /// <summary>
         /// Returns total byte size used by cached filters. </summary>
         public virtual long SizeInBytes()
-        {
-            // Sync only to pull the current set of values:
-            lock (_cache)
-            {
-                IList<DocIdSet> docIdSets = new List<DocIdSet>(_cache.Values);
-                return docIdSets.Sum(dis => RamUsageEstimator.SizeOf(dis));
-            }
+        {            
+            IList<DocIdSet> docIdSets = new List<DocIdSet>(_cache.Values);
+            return docIdSets.Sum(dis => RamUsageEstimator.SizeOf(dis));
         }
     }
 }
\ No newline at end of file