You are viewing a plain text version of this content. The canonical link for it is here.
Posted to pylucene-commits@lucene.apache.org by va...@apache.org on 2010/07/12 21:33:47 UTC
svn commit: r963446 [2/2] - in /lucene/pylucene/branches/python_3: ./
samples/LuceneInAction/ samples/LuceneInAction/lia/advsearching/
samples/LuceneInAction/lia/analysis/
samples/LuceneInAction/lia/analysis/keyword/
samples/LuceneInAction/lia/analysis...
Modified: lucene/pylucene/branches/python_3/test/test_Sort.py
URL: http://svn.apache.org/viewvc/lucene/pylucene/branches/python_3/test/test_Sort.py?rev=963446&r1=963445&r2=963446&view=diff
==============================================================================
--- lucene/pylucene/branches/python_3/test/test_Sort.py (original)
+++ lucene/pylucene/branches/python_3/test/test_Sort.py Mon Jul 12 19:33:45 2010
@@ -14,7 +14,7 @@
import math
-from itertools import izip
+
from random import randint
from unittest import TestCase, main
from lucene import *
@@ -34,15 +34,15 @@ class SortTestCase(TestCase):
self.data = [
# tracer contents int float string custom i18n long double, short, byte, custom parser encoding'
- [ "A", "x a", "5", "4f", "c", "A-3", u"p\u00EAche", "10", "-4.0", "3", "126", "J" ],
+ [ "A", "x a", "5", "4f", "c", "A-3", "p\u00EAche", "10", "-4.0", "3", "126", "J" ],
[ "B", "y a", "5", "3.4028235E38", "i", "B-10", "HAT", "1000000000", "40.0", "24", "1", "I" ],
- [ "C", "x a b c", "2147483647", "1.0", "j", "A-2", u"p\u00E9ch\u00E9", "99999999", "40.00002343", "125", "15", "H" ],
+ [ "C", "x a b c", "2147483647", "1.0", "j", "A-2", "p\u00E9ch\u00E9", "99999999", "40.00002343", "125", "15", "H" ],
[ "D", "y a b c", "-1", "0.0f", "a", "C-0", "HUT", str(Long.MAX_VALUE), str(Double.MIN_VALUE), str(Short.MIN_VALUE), str(Byte.MIN_VALUE), "G" ],
[ "E", "x a b c d", "5", "2f", "h", "B-8", "peach", str(Long.MIN_VALUE), str(Double.MAX_VALUE), str(Short.MAX_VALUE), str(Byte.MAX_VALUE), "F" ],
- [ "F", "y a b c d", "2", "3.14159f", "g", "B-1", u"H\u00C5T", "-44", "343.034435444", "-3", "0", "E" ],
+ [ "F", "y a b c d", "2", "3.14159f", "g", "B-1", "H\u00C5T", "-44", "343.034435444", "-3", "0", "E" ],
[ "G", "x a b c d", "3", "-1.0", "f", "C-100", "sin", "323254543543", "4.043544", "5", "100", "D" ],
- [ "H", "y a b c d", "0", "1.4E-45", "e", "C-88", u"H\u00D8T", "1023423423005", "4.043545", "10", "-50", "C" ],
- [ "I", "x a b c d e f", "-2147483648", "1.0e+0", "d", "A-10", u"s\u00EDn", "332422459999", "4.043546", "-340", "51", "B" ],
+ [ "H", "y a b c d", "0", "1.4E-45", "e", "C-88", "H\u00D8T", "1023423423005", "4.043545", "10", "-50", "C" ],
+ [ "I", "x a b c d e f", "-2147483648", "1.0e+0", "d", "A-10", "s\u00EDn", "332422459999", "4.043546", "-340", "51", "B" ],
[ "J", "y a b c d e f", "4", ".5", "b", "C-7", "HOT", "34334543543", "4.0000220343", "300", "2", "A" ],
[ "W", "g", "1", None, None, None, None, None, None, None, None, None ],
[ "X", "g", "1", "0.1", None, None, None, None, None, None, None, None ],
@@ -58,7 +58,7 @@ class SortTestCase(TestCase):
writer.setMaxBufferedDocs(2)
writer.setMergeFactor(1000)
- for i in xrange(len(self.data)):
+ for i in range(len(self.data)):
if (i % 2 == 0 and even) or (i % 2 == 1 and odd):
doc = Document()
doc.add(Field("tracer", self.data[i][0], Field.Store.YES,
@@ -115,7 +115,7 @@ class SortTestCase(TestCase):
writer.setMaxBufferedDocs(4)
writer.setMergeFactor(97)
- for i in xrange(NUM_STRINGS):
+ for i in range(NUM_STRINGS):
doc = Document()
num = self.getRandomCharString(self.getRandomNumber(2, 8), 48, 52)
doc.add(Field("tracer", num, Field.Store.YES, Field.Index.NO))
@@ -139,7 +139,7 @@ class SortTestCase(TestCase):
def getRandomNumberString(self, num, low, high):
- return ''.join([self.getRandomNumber(low, high) for i in xrange(num)])
+ return ''.join([self.getRandomNumber(low, high) for i in range(num)])
def getRandomCharString(self, num):
@@ -148,7 +148,7 @@ class SortTestCase(TestCase):
def getRandomCharString(self, num, start, end):
return ''.join([chr(self.getRandomNumber(start, end))
- for i in xrange(num)])
+ for i in range(num)])
def getRandomNumber(self, low, high):
@@ -255,22 +255,22 @@ class SortTestCase(TestCase):
doc2 = searcher.doc(scoreDoc.doc)
v = doc2.getValues("tracer")
v2 = doc2.getValues("tracer2")
- for _v, _v2 in izip(v, v2):
+ for _v, _v2 in zip(v, v2):
if last is not None:
- _cmp = cmp(_v, last)
+ _cmp = (_v > last) - (_v < last)
if _cmp < 0: # ensure first field is in order
fail = True
- print "fail:", _v, "<", last
+ print("fail:", _v, "<", last)
if _cmp == 0: # ensure second field is in reverse order
- _cmp = cmp(_v2, lastSub)
+ _cmp = (_v2 > lastSub) - (_v2 < lastSub)
if _cmp > 0:
fail = True
- print "rev field fail:", _v2, ">", lastSub
+ print("rev field fail:", _v2, ">", lastSub)
elif _cmp == 0: # ensure docid is in order
if scoreDoc.doc < lastDocId:
fail = True
- print "doc fail:", scoreDoc.doc, ">", lastDocId
+ print("doc fail:", scoreDoc.doc, ">", lastDocId)
last = _v
lastSub = _v2
@@ -278,7 +278,7 @@ class SortTestCase(TestCase):
buff.append(_v + "(" + _v2 + ")(" + str(scoreDoc.doc) + ") ")
if fail:
- print "topn field1(field2)(docID):", ''.join(buff)
+ print("topn field1(field2)(docID):", ''.join(buff))
self.assert_(not fail, "Found sort results out of order")
@@ -305,7 +305,7 @@ class SortTestCase(TestCase):
class longParser(PythonLongParser):
def parseLong(_self, val):
- return (val.bytes[0] - ord('A')) * 1234567890L
+ return (val.bytes[0] - ord('A')) * 1234567890
class doubleParser(PythonDoubleParser):
def parseDouble(_self, val):
@@ -717,7 +717,7 @@ class SortTestCase(TestCase):
self.full.search(q, tdc)
sds = tdc.topDocs().scoreDocs
- for i in xrange(1, len(sds)):
+ for i in range(1, len(sds)):
self.assert_(sds[i].doc != sds[i - 1].doc)
def testSortWithoutScoreTracking(self):
@@ -818,7 +818,7 @@ class SortTestCase(TestCase):
bq.setMinimumNumberShouldMatch(1)
for sort in sorts:
- for tfcOption, actualTFCClass in izip(tfcOptions,
+ for tfcOption, actualTFCClass in zip(tfcOptions,
actualTFCClasses):
tdc = TopFieldCollector.create(sort, 10, tfcOption[0],
tfcOption[1], tfcOption[2],
@@ -965,7 +965,7 @@ class SortTestCase(TestCase):
"""
self.assertEquals(len(m1), len(m2))
- for key in m1.iterkeys():
+ for key in m1.keys():
self.assertEquals(m1[key], m2[key], 1e-6)
def getName(self):
Modified: lucene/pylucene/branches/python_3/test/test_StopWords.py
URL: http://svn.apache.org/viewvc/lucene/pylucene/branches/python_3/test/test_StopWords.py?rev=963446&r1=963445&r2=963446&view=diff
==============================================================================
--- lucene/pylucene/branches/python_3/test/test_StopWords.py (original)
+++ lucene/pylucene/branches/python_3/test/test_StopWords.py Mon Jul 12 19:33:45 2010
@@ -34,7 +34,7 @@ class StopWordsTestCase(TestCase):
try:
result = StandardTokenizer(Version.LUCENE_CURRENT, self.reader)
result = StopFilter(True, result, self.stop_set)
- except Exception, e:
+ except Exception as e:
self.fail(str(e))
Modified: lucene/pylucene/branches/python_3/test/test_TermRangeFilter.py
URL: http://svn.apache.org/viewvc/lucene/pylucene/branches/python_3/test/test_TermRangeFilter.py?rev=963446&r1=963445&r2=963446&view=diff
==============================================================================
--- lucene/pylucene/branches/python_3/test/test_TermRangeFilter.py (original)
+++ lucene/pylucene/branches/python_3/test/test_TermRangeFilter.py Mon Jul 12 19:33:45 2010
@@ -393,7 +393,7 @@ class TestTermRangeFilter(BaseTestRangeF
writer = IndexWriter(farsiIndex, SimpleAnalyzer(), True,
IndexWriter.MaxFieldLength.LIMITED)
doc = Document()
- doc.add(Field("content", u"\u0633\u0627\u0628",
+ doc.add(Field("content", "\u0633\u0627\u0628",
Field.Store.YES, Field.Index.NOT_ANALYZED))
doc.add(Field("body", "body",
Field.Store.YES, Field.Index.NOT_ANALYZED))
@@ -416,10 +416,10 @@ class TestTermRangeFilter(BaseTestRangeF
# single index Term below should NOT be returned by a
# TermRangeFilter with a Farsi Collator (or an Arabic one for the
# case when Farsi is not supported).
- numHits = search.search(q, TermRangeFilter("content", u"\u062F", u"\u0698", True, True, collator), 1000).totalHits
+ numHits = search.search(q, TermRangeFilter("content", "\u062F", "\u0698", True, True, collator), 1000).totalHits
self.assertEqual(0, numHits, "The index Term should not be included.")
- numHits = search.search(q, TermRangeFilter("content", u"\u0633", u"\u0638", True, True, collator), 1000).totalHits
+ numHits = search.search(q, TermRangeFilter("content", "\u0633", "\u0638", True, True, collator), 1000).totalHits
self.assertEqual(1, numHits, "The index Term should be included.")
search.close()
@@ -432,7 +432,7 @@ class TestTermRangeFilter(BaseTestRangeF
# Danish collation orders the words below in the given order
# (example taken from TestSort.testInternationalSort() ).
- words = [u"H\u00D8T", u"H\u00C5T", "MAND"]
+ words = ["H\u00D8T", "H\u00C5T", "MAND"]
for word in words:
doc = Document()
doc.add(Field("content", word, Field.Store.YES,
@@ -454,10 +454,10 @@ class TestTermRangeFilter(BaseTestRangeF
# Unicode order would not include "H\u00C5T" in [ "H\u00D8T", "MAND" ],
# but Danish collation does.
- numHits = search.search(q, TermRangeFilter("content", u"H\u00D8T", "MAND", False, False, collator), 1000).totalHits
+ numHits = search.search(q, TermRangeFilter("content", "H\u00D8T", "MAND", False, False, collator), 1000).totalHits
self.assertEqual(1, numHits, "The index Term should be included.")
- numHits = search.search(q, TermRangeFilter("content", u"H\u00C5T", "MAND", False, False, collator), 1000).totalHits
+ numHits = search.search(q, TermRangeFilter("content", "H\u00C5T", "MAND", False, False, collator), 1000).totalHits
self.assertEqual(0, numHits, "The index Term should not be included.")
search.close()
Modified: lucene/pylucene/branches/python_3/test/test_ThaiAnalyzer.py
URL: http://svn.apache.org/viewvc/lucene/pylucene/branches/python_3/test/test_ThaiAnalyzer.py?rev=963446&r1=963445&r2=963446&view=diff
==============================================================================
--- lucene/pylucene/branches/python_3/test/test_ThaiAnalyzer.py (original)
+++ lucene/pylucene/branches/python_3/test/test_ThaiAnalyzer.py Mon Jul 12 19:33:45 2010
@@ -24,22 +24,22 @@ class ThaiAnalyzerTestCase(BaseTokenStre
analyzer = ThaiAnalyzer(Version.LUCENE_CURRENT)
- self._assertAnalyzesTo(analyzer, u"", [])
+ self._assertAnalyzesTo(analyzer, "", [])
self._assertAnalyzesTo(analyzer,
- u"à¸à¸²à¸£à¸à¸µà¹à¹à¸à¹à¸à¹à¸à¸à¹à¸ªà¸à¸à¸§à¹à¸²à¸à¸²à¸à¸à¸µ",
- [ u"à¸à¸²à¸£", u"à¸à¸µà¹", u"à¹à¸à¹", u"à¸à¹à¸à¸",
- u"à¹à¸ªà¸à¸", u"วà¹à¸²", u"à¸à¸²à¸", u"à¸à¸µ" ])
+ "à¸à¸²à¸£à¸à¸µà¹à¹à¸à¹à¸à¹à¸à¸à¹à¸ªà¸à¸à¸§à¹à¸²à¸à¸²à¸à¸à¸µ",
+ [ "à¸à¸²à¸£", "à¸à¸µà¹", "à¹à¸à¹", "à¸à¹à¸à¸",
+ "à¹à¸ªà¸à¸", "วà¹à¸²", "à¸à¸²à¸", "à¸à¸µ" ])
self._assertAnalyzesTo(analyzer,
- u"à¸à¸£à¸´à¸©à¸±à¸à¸à¸·à¹à¸ XY&Z - à¸à¸¸à¸¢à¸à¸±à¸ xyz@demo.com",
- [ u"à¸à¸£à¸´à¸©à¸±à¸", u"à¸à¸·à¹à¸", u"xy&z", u"à¸à¸¸à¸¢", u"à¸à¸±à¸", u"xyz@demo.com" ])
+ "à¸à¸£à¸´à¸©à¸±à¸à¸à¸·à¹à¸ XY&Z - à¸à¸¸à¸¢à¸à¸±à¸ xyz@demo.com",
+ [ "à¸à¸£à¸´à¸©à¸±à¸", "à¸à¸·à¹à¸", "xy&z", "à¸à¸¸à¸¢", "à¸à¸±à¸", "xyz@demo.com" ])
# English stop words
self._assertAnalyzesTo(analyzer,
- u"à¸à¸£à¸°à¹à¸¢à¸à¸§à¹à¸² The quick brown fox jumped over the lazy dogs",
- [ u"à¸à¸£à¸°à¹à¸¢à¸", u"วà¹à¸²", u"quick", u"brown", u"fox",
- u"jumped", u"over", u"lazy", u"dogs" ])
+ "à¸à¸£à¸°à¹à¸¢à¸à¸§à¹à¸² The quick brown fox jumped over the lazy dogs",
+ [ "à¸à¸£à¸°à¹à¸¢à¸", "วà¹à¸²", "quick", "brown", "fox",
+ "jumped", "over", "lazy", "dogs" ])
if __name__ == "__main__":
Modified: lucene/pylucene/branches/python_3/test/test_bug1564.py
URL: http://svn.apache.org/viewvc/lucene/pylucene/branches/python_3/test/test_bug1564.py?rev=963446&r1=963445&r2=963446&view=diff
==============================================================================
--- lucene/pylucene/branches/python_3/test/test_bug1564.py (original)
+++ lucene/pylucene/branches/python_3/test/test_bug1564.py Mon Jul 12 19:33:45 2010
@@ -25,7 +25,7 @@ class Test_Bug1564(unittest.TestCase):
writer = IndexWriter(self.store, self.analyzer, True,
IndexWriter.MaxFieldLength.LIMITED)
doc = Document()
- doc.add(Field('all', u'windowpane beplaster rapacious \
+ doc.add(Field('all', 'windowpane beplaster rapacious \
catatonia gauntlet wynn depressible swede pick dressmake supreme \
jeremy plumb theoretic bureaucracy causation chartres equipoise \
dispersible careen heard',