You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by rm...@apache.org on 2011/08/30 01:13:23 UTC
svn commit: r1163047 [7/15] - in /lucene/dev/branches/flexscoring: ./
dev-tools/idea/lucene/contrib/ lucene/ lucene/contrib/
lucene/contrib/demo/src/java/org/apache/lucene/demo/
lucene/contrib/demo/src/java/org/apache/lucene/demo/xmlparser/ lucene/cont...
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java Mon Aug 29 23:13:10 2011
@@ -20,11 +20,12 @@ package org.apache.lucene.index;
import java.io.IOException;
import org.apache.lucene.analysis.MockAnalyzer;
+import org.apache.lucene.document.BinaryField;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
-import org.apache.lucene.document.Field.Index;
-import org.apache.lucene.document.Field.Store;
-import org.apache.lucene.document.Field.TermVector;
+import org.apache.lucene.document.FieldType;
+import org.apache.lucene.document.StringField;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.store.Directory;
import org.apache.lucene.util.LuceneTestCase;
import org.junit.Test;
@@ -38,8 +39,8 @@ public class TestConsistentFieldNumbers
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(NoMergePolicy.COMPOUND_FILES));
Document d1 = new Document();
- d1.add(new Field("f1", "first field", Store.YES, Index.ANALYZED, TermVector.NO));
- d1.add(new Field("f2", "second field", Store.YES, Index.ANALYZED, TermVector.NO));
+ d1.add(new Field("f1", StringField.TYPE_STORED, "first field"));
+ d1.add(new Field("f2", StringField.TYPE_STORED, "second field"));
writer.addDocument(d1);
if (i == 1) {
@@ -50,10 +51,12 @@ public class TestConsistentFieldNumbers
}
Document d2 = new Document();
- d2.add(new Field("f2", "second field", Store.YES, Index.ANALYZED, TermVector.NO));
- d2.add(new Field("f1", "first field", Store.YES, Index.ANALYZED, TermVector.YES));
- d2.add(new Field("f3", "third field", Store.YES, Index.ANALYZED, TermVector.NO));
- d2.add(new Field("f4", "fourth field", Store.YES, Index.ANALYZED, TermVector.NO));
+ FieldType customType2 = new FieldType(TextField.TYPE_STORED);
+ customType2.setStoreTermVectors(true);
+ d2.add(new TextField("f2", "second field"));
+ d2.add(new Field("f1", customType2, "first field"));
+ d2.add(new TextField("f3", "third field"));
+ d2.add(new TextField("f4", "fourth field"));
writer.addDocument(d2);
writer.close();
@@ -99,18 +102,20 @@ public class TestConsistentFieldNumbers
IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(NoMergePolicy.COMPOUND_FILES));
Document d1 = new Document();
- d1.add(new Field("f1", "first field", Store.YES, Index.ANALYZED, TermVector.NO));
- d1.add(new Field("f2", "second field", Store.YES, Index.ANALYZED, TermVector.NO));
+ d1.add(new Field("f1", TextField.TYPE_STORED, "first field"));
+ d1.add(new Field("f2", TextField.TYPE_STORED, "second field"));
writer.addDocument(d1);
writer.close();
writer = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(NoMergePolicy.COMPOUND_FILES));
Document d2 = new Document();
- d2.add(new Field("f2", "second field", Store.YES, Index.ANALYZED, TermVector.NO));
- d2.add(new Field("f1", "first field", Store.YES, Index.ANALYZED, TermVector.YES));
- d2.add(new Field("f3", "third field", Store.YES, Index.ANALYZED, TermVector.NO));
- d2.add(new Field("f4", "fourth field", Store.YES, Index.ANALYZED, TermVector.NO));
+ FieldType customType2 = new FieldType(TextField.TYPE_STORED);
+ customType2.setStoreTermVectors(true);
+ d2.add(new Field("f2", TextField.TYPE_STORED, "second field"));
+ d2.add(new Field("f1", customType2, "first field"));
+ d2.add(new Field("f3", TextField.TYPE_STORED, "third field"));
+ d2.add(new Field("f4", TextField.TYPE_STORED, "fourth field"));
writer.addDocument(d2);
writer.close();
@@ -163,10 +168,8 @@ public class TestConsistentFieldNumbers
TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(
NoMergePolicy.NO_COMPOUND_FILES));
Document d = new Document();
- d.add(new Field("f1", "d1 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d1 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d1 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d1 second field"));
writer.addDocument(d);
writer.close();
SegmentInfos sis = new SegmentInfos();
@@ -185,9 +188,8 @@ public class TestConsistentFieldNumbers
random.nextBoolean() ? NoMergePolicy.NO_COMPOUND_FILES
: NoMergePolicy.COMPOUND_FILES));
Document d = new Document();
- d.add(new Field("f1", "d2 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d2 first field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3 }));
writer.addDocument(d);
writer.close();
SegmentInfos sis = new SegmentInfos();
@@ -210,11 +212,9 @@ public class TestConsistentFieldNumbers
random.nextBoolean() ? NoMergePolicy.NO_COMPOUND_FILES
: NoMergePolicy.COMPOUND_FILES));
Document d = new Document();
- d.add(new Field("f1", "d3 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d3 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3, 4, 5 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d3 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d3 second field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3, 4, 5 }));
writer.addDocument(d);
writer.close();
SegmentInfos sis = new SegmentInfos();
@@ -303,10 +303,10 @@ public class TestConsistentFieldNumbers
for (FieldInfo fi : fis) {
Field expected = getField(Integer.parseInt(fi.name));
- assertEquals(expected.isIndexed(), fi.isIndexed);
- assertEquals(expected.isTermVectorStored(), fi.storeTermVector);
- assertEquals(expected.isStorePositionWithTermVector(), fi.storePositionWithTermVector);
- assertEquals(expected.isStoreOffsetWithTermVector(), fi.storeOffsetWithTermVector);
+ assertEquals(expected.indexed(), fi.isIndexed);
+ assertEquals(expected.storeTermVectors(), fi.storeTermVector);
+ assertEquals(expected.storeTermVectorPositions(), fi.storePositionWithTermVector);
+ assertEquals(expected.storeTermVectorOffsets(), fi.storeOffsetWithTermVector);
}
}
@@ -316,23 +316,91 @@ public class TestConsistentFieldNumbers
private Field getField(int number) {
int mode = number % 16;
String fieldName = "" + number;
+ FieldType customType = new FieldType(TextField.TYPE_STORED);
+
+ FieldType customType2 = new FieldType(TextField.TYPE_STORED);
+ customType2.setTokenized(false);
+
+ FieldType customType3 = new FieldType(TextField.TYPE_UNSTORED);
+ customType3.setTokenized(false);
+
+ FieldType customType4 = new FieldType(TextField.TYPE_UNSTORED);
+ customType4.setTokenized(false);
+ customType4.setStoreTermVectors(true);
+ customType4.setStoreTermVectorOffsets(true);
+
+ FieldType customType5 = new FieldType(TextField.TYPE_UNSTORED);
+ customType5.setStoreTermVectors(true);
+ customType5.setStoreTermVectorOffsets(true);
+
+ FieldType customType6 = new FieldType(TextField.TYPE_STORED);
+ customType6.setTokenized(false);
+ customType6.setStoreTermVectors(true);
+ customType6.setStoreTermVectorOffsets(true);
+
+ FieldType customType7 = new FieldType(TextField.TYPE_UNSTORED);
+ customType7.setTokenized(false);
+ customType7.setStoreTermVectors(true);
+ customType7.setStoreTermVectorOffsets(true);
+
+ FieldType customType8 = new FieldType(TextField.TYPE_STORED);
+ customType8.setTokenized(false);
+ customType8.setStoreTermVectors(true);
+ customType8.setStoreTermVectorPositions(true);
+
+ FieldType customType9 = new FieldType(TextField.TYPE_UNSTORED);
+ customType9.setStoreTermVectors(true);
+ customType9.setStoreTermVectorPositions(true);
+
+ FieldType customType10 = new FieldType(TextField.TYPE_STORED);
+ customType10.setTokenized(false);
+ customType10.setStoreTermVectors(true);
+ customType10.setStoreTermVectorPositions(true);
+
+ FieldType customType11 = new FieldType(TextField.TYPE_UNSTORED);
+ customType11.setTokenized(false);
+ customType11.setStoreTermVectors(true);
+ customType11.setStoreTermVectorPositions(true);
+
+ FieldType customType12 = new FieldType(TextField.TYPE_STORED);
+ customType12.setStoreTermVectors(true);
+ customType12.setStoreTermVectorOffsets(true);
+ customType12.setStoreTermVectorPositions(true);
+
+ FieldType customType13 = new FieldType(TextField.TYPE_UNSTORED);
+ customType13.setStoreTermVectors(true);
+ customType13.setStoreTermVectorOffsets(true);
+ customType13.setStoreTermVectorPositions(true);
+
+ FieldType customType14 = new FieldType(TextField.TYPE_STORED);
+ customType14.setTokenized(false);
+ customType14.setStoreTermVectors(true);
+ customType14.setStoreTermVectorOffsets(true);
+ customType14.setStoreTermVectorPositions(true);
+
+ FieldType customType15 = new FieldType(TextField.TYPE_UNSTORED);
+ customType15.setTokenized(false);
+ customType15.setStoreTermVectors(true);
+ customType15.setStoreTermVectorOffsets(true);
+ customType15.setStoreTermVectorPositions(true);
+
switch (mode) {
- case 0: return new Field(fieldName, "some text", Store.YES, Index.ANALYZED, TermVector.NO);
- case 1: return new Field(fieldName, "some text", Store.NO, Index.ANALYZED, TermVector.NO);
- case 2: return new Field(fieldName, "some text", Store.YES, Index.NOT_ANALYZED, TermVector.NO);
- case 3: return new Field(fieldName, "some text", Store.NO, Index.NOT_ANALYZED, TermVector.NO);
- case 4: return new Field(fieldName, "some text", Store.YES, Index.ANALYZED, TermVector.WITH_OFFSETS);
- case 5: return new Field(fieldName, "some text", Store.NO, Index.ANALYZED, TermVector.WITH_OFFSETS);
- case 6: return new Field(fieldName, "some text", Store.YES, Index.NOT_ANALYZED, TermVector.WITH_OFFSETS);
- case 7: return new Field(fieldName, "some text", Store.NO, Index.NOT_ANALYZED, TermVector.WITH_OFFSETS);
- case 8: return new Field(fieldName, "some text", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS);
- case 9: return new Field(fieldName, "some text", Store.NO, Index.ANALYZED, TermVector.WITH_POSITIONS);
- case 10: return new Field(fieldName, "some text", Store.YES, Index.NOT_ANALYZED, TermVector.WITH_POSITIONS);
- case 11: return new Field(fieldName, "some text", Store.NO, Index.NOT_ANALYZED, TermVector.WITH_POSITIONS);
- case 12: return new Field(fieldName, "some text", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS);
- case 13: return new Field(fieldName, "some text", Store.NO, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS);
- case 14: return new Field(fieldName, "some text", Store.YES, Index.NOT_ANALYZED, TermVector.WITH_POSITIONS_OFFSETS);
- case 15: return new Field(fieldName, "some text", Store.NO, Index.NOT_ANALYZED, TermVector.WITH_POSITIONS_OFFSETS);
+ case 0: return new Field(fieldName, customType, "some text");
+ case 1: return new TextField(fieldName, "some text");
+ case 2: return new Field(fieldName, customType2, "some text");
+ case 3: return new Field(fieldName, customType3, "some text");
+ case 4: return new Field(fieldName, customType4, "some text");
+ case 5: return new Field(fieldName, customType5, "some text");
+ case 6: return new Field(fieldName, customType6, "some text");
+ case 7: return new Field(fieldName, customType7, "some text");
+ case 8: return new Field(fieldName, customType8, "some text");
+ case 9: return new Field(fieldName, customType9, "some text");
+ case 10: return new Field(fieldName, customType10, "some text");
+ case 11: return new Field(fieldName, customType11, "some text");
+ case 12: return new Field(fieldName, customType12, "some text");
+ case 13: return new Field(fieldName, customType13, "some text");
+ case 14: return new Field(fieldName, customType14, "some text");
+ case 15: return new Field(fieldName, customType15, "some text");
default: return null;
}
}
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestCrash.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestCrash.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestCrash.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestCrash.java Mon Aug 29 23:13:10 2011
@@ -25,7 +25,7 @@ import org.apache.lucene.store.MockDirec
import org.apache.lucene.store.NoLockFactory;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
+import org.apache.lucene.document.TextField;
public class TestCrash extends LuceneTestCase {
@@ -44,8 +44,8 @@ public class TestCrash extends LuceneTes
}
Document doc = new Document();
- doc.add(newField("content", "aaa", Field.Store.YES, Field.Index.ANALYZED));
- doc.add(newField("id", "0", Field.Store.YES, Field.Index.ANALYZED));
+ doc.add(newField("content", "aaa", TextField.TYPE_UNSTORED));
+ doc.add(newField("id", "0", TextField.TYPE_UNSTORED));
for(int i=0;i<157;i++)
writer.addDocument(doc);
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java Mon Aug 29 23:13:10 2011
@@ -25,7 +25,7 @@ import java.util.Collection;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.IndexWriterConfig.OpenMode;
import org.apache.lucene.search.IndexSearcher;
import org.apache.lucene.search.Query;
@@ -853,7 +853,7 @@ public class TestDeletionPolicy extends
private void addDoc(IndexWriter writer) throws IOException
{
Document doc = new Document();
- doc.add(newField("content", "aaa", Field.Store.NO, Field.Index.ANALYZED));
+ doc.add(newField("content", "aaa", TextField.TYPE_UNSTORED));
writer.addDocument(doc);
}
}
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java Mon Aug 29 23:13:10 2011
@@ -21,7 +21,7 @@ import org.apache.lucene.util.LuceneTest
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.IndexWriterConfig.OpenMode;
import org.apache.lucene.store.Directory;
import org.apache.lucene.util.BytesRef;
@@ -202,7 +202,7 @@ public class TestDirectoryReader extends
new MockAnalyzer(random)).setOpenMode(
create ? OpenMode.CREATE : OpenMode.APPEND));
Document doc = new Document();
- doc.add(newField("body", s, Field.Store.YES, Field.Index.ANALYZED));
+ doc.add(newField("body", s, TextField.TYPE_UNSTORED));
iw.addDocument(doc);
iw.close();
}
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDoc.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDoc.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDoc.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDoc.java Mon Aug 29 23:13:10 2011
@@ -31,14 +31,13 @@ import junit.textui.TestRunner;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.IndexWriterConfig.OpenMode;
import org.apache.lucene.search.DocIdSetIterator;
import org.apache.lucene.store.Directory;
import org.apache.lucene.store.IOContext;
import org.apache.lucene.util.LuceneTestCase;
import org.apache.lucene.util._TestUtil;
-import org.apache.lucene.index.codecs.CodecProvider;
/** JUnit adaptation of an older test case DocTest. */
@@ -185,7 +184,7 @@ public class TestDoc extends LuceneTestC
{
File file = new File(workDir, fileName);
Document doc = new Document();
- doc.add(new Field("contents", new FileReader(file)));
+ doc.add(new TextField("contents", new FileReader(file)));
writer.addDocument(doc);
writer.commit();
return writer.newestSegment();
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java Mon Aug 29 23:13:10 2011
@@ -28,6 +28,8 @@ import org.apache.lucene.analysis.MockAn
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
import org.apache.lucene.document.NumericField;
+import org.apache.lucene.document.StringField;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.DocTermOrds.TermOrdsIterator;
import org.apache.lucene.index.codecs.BlockTermsReader;
import org.apache.lucene.index.codecs.BlockTermsWriter;
@@ -66,7 +68,7 @@ public class TestDocTermOrds extends Luc
Directory dir = newDirectory();
final RandomIndexWriter w = new RandomIndexWriter(random, dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy()));
Document doc = new Document();
- Field field = newField("field", "", Field.Index.ANALYZED);
+ Field field = newField("field", "", TextField.TYPE_UNSTORED);
doc.add(field);
field.setValue("a b c");
w.addDocument(doc);
@@ -281,7 +283,7 @@ public class TestDocTermOrds extends Luc
}
for(int ord : ordsForDocSet) {
ordsForDoc[upto++] = ord;
- Field field = newField("field", termsArray[ord].utf8ToString(), Field.Index.NOT_ANALYZED_NO_NORMS);
+ Field field = newField("field", termsArray[ord].utf8ToString(), StringField.TYPE_UNSTORED);
if (VERBOSE) {
System.out.println(" f=" + termsArray[ord].utf8ToString());
}
@@ -384,7 +386,7 @@ public class TestDocTermOrds extends Luc
}
for(int ord : ordsForDocSet) {
ordsForDoc[upto++] = ord;
- Field field = newField("field", termsArray[ord].utf8ToString(), Field.Index.NOT_ANALYZED_NO_NORMS);
+ Field field = newField("field", termsArray[ord].utf8ToString(), StringField.TYPE_UNSTORED);
if (VERBOSE) {
System.out.println(" f=" + termsArray[ord].utf8ToString());
}
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocsAndPositions.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocsAndPositions.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocsAndPositions.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocsAndPositions.java Mon Aug 29 23:13:10 2011
@@ -22,7 +22,8 @@ import java.util.Arrays;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
+import org.apache.lucene.document.FieldType;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.IndexReader.AtomicReaderContext;
import org.apache.lucene.index.IndexReader.ReaderContext;
import org.apache.lucene.store.Directory;
@@ -49,9 +50,11 @@ public class TestDocsAndPositions extend
newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
for (int i = 0; i < 39; i++) {
Document doc = new Document();
+ FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
+ customType.setOmitNorms(true);
doc.add(newField(fieldName, "1 2 3 4 5 6 7 8 9 10 "
+ "1 2 3 4 5 6 7 8 9 10 " + "1 2 3 4 5 6 7 8 9 10 "
- + "1 2 3 4 5 6 7 8 9 10", Field.Store.NO, Field.Index.ANALYZED_NO_NORMS));
+ + "1 2 3 4 5 6 7 8 9 10", customType));
writer.addDocument(doc);
}
IndexReader reader = writer.getReader();
@@ -117,6 +120,8 @@ public class TestDocsAndPositions extend
int max = 1051;
int term = random.nextInt(max);
Integer[][] positionsInDoc = new Integer[numDocs][];
+ FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
+ customType.setOmitNorms(true);
for (int i = 0; i < numDocs; i++) {
Document doc = new Document();
ArrayList<Integer> positions = new ArrayList<Integer>();
@@ -133,8 +138,7 @@ public class TestDocsAndPositions extend
builder.append(term);
positions.add(num);
}
- doc.add(newField(fieldName, builder.toString(), Field.Store.NO,
- Field.Index.ANALYZED_NO_NORMS));
+ doc.add(newField(fieldName, builder.toString(), customType));
positionsInDoc[i] = positions.toArray(new Integer[0]);
writer.addDocument(doc);
}
@@ -200,6 +204,8 @@ public class TestDocsAndPositions extend
int max = 15678;
int term = random.nextInt(max);
int[] freqInDoc = new int[numDocs];
+ FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
+ customType.setOmitNorms(true);
for (int i = 0; i < numDocs; i++) {
Document doc = new Document();
StringBuilder builder = new StringBuilder();
@@ -210,8 +216,7 @@ public class TestDocsAndPositions extend
freqInDoc[i]++;
}
}
- doc.add(newField(fieldName, builder.toString(), Field.Store.NO,
- Field.Index.ANALYZED_NO_NORMS));
+ doc.add(newField(fieldName, builder.toString(), customType));
writer.addDocument(doc);
}
@@ -276,6 +281,8 @@ public class TestDocsAndPositions extend
RandomIndexWriter writer = new RandomIndexWriter(random, dir,
newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
int howMany = 1000;
+ FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
+ customType.setOmitNorms(true);
for (int i = 0; i < 39; i++) {
Document doc = new Document();
StringBuilder builder = new StringBuilder();
@@ -286,8 +293,7 @@ public class TestDocsAndPositions extend
builder.append("odd ");
}
}
- doc.add(newField(fieldName, builder.toString(), Field.Store.NO,
- Field.Index.ANALYZED_NO_NORMS));
+ doc.add(newField(fieldName, builder.toString(), customType));
writer.addDocument(doc);
}
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocumentWriter.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocumentWriter.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocumentWriter.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestDocumentWriter.java Mon Aug 29 23:13:10 2011
@@ -30,14 +30,11 @@ import org.apache.lucene.analysis.tokena
import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
-import org.apache.lucene.document.Field.Index;
-import org.apache.lucene.document.Field.Store;
-import org.apache.lucene.document.Field.TermVector;
-import org.apache.lucene.document.Fieldable;
+import org.apache.lucene.document.FieldType;
+import org.apache.lucene.document.StringField;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.FieldInfo.IndexOptions;
import org.apache.lucene.store.Directory;
-import org.apache.lucene.store.IOContext;
-import org.apache.lucene.store.IOContext.Context;
import org.apache.lucene.util.AttributeSource;
import org.apache.lucene.util.BytesRef;
import org.apache.lucene.util.LuceneTestCase;
@@ -77,15 +74,15 @@ public class TestDocumentWriter extends
assertTrue(doc != null);
//System.out.println("Document: " + doc);
- Fieldable [] fields = doc.getFields("textField2");
+ IndexableField [] fields = doc.getFields("textField2");
assertTrue(fields != null && fields.length == 1);
assertTrue(fields[0].stringValue().equals(DocHelper.FIELD_2_TEXT));
- assertTrue(fields[0].isTermVectorStored());
+ assertTrue(fields[0].storeTermVectors());
fields = doc.getFields("textField1");
assertTrue(fields != null && fields.length == 1);
assertTrue(fields[0].stringValue().equals(DocHelper.FIELD_1_TEXT));
- assertFalse(fields[0].isTermVectorStored());
+ assertFalse(fields[0].storeTermVectors());
fields = doc.getFields("keyField");
assertTrue(fields != null && fields.length == 1);
@@ -125,8 +122,8 @@ public class TestDocumentWriter extends
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer));
Document doc = new Document();
- doc.add(newField("repeated", "repeated one", Field.Store.YES, Field.Index.ANALYZED));
- doc.add(newField("repeated", "repeated two", Field.Store.YES, Field.Index.ANALYZED));
+ doc.add(newField("repeated", "repeated one", TextField.TYPE_STORED));
+ doc.add(newField("repeated", "repeated two", TextField.TYPE_STORED));
writer.addDocument(doc);
writer.commit();
@@ -197,7 +194,7 @@ public class TestDocumentWriter extends
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer));
Document doc = new Document();
- doc.add(newField("f1", "a 5 a a", Field.Store.YES, Field.Index.ANALYZED));
+ doc.add(newField("f1", "a 5 a a", TextField.TYPE_STORED));
writer.addDocument(doc);
writer.commit();
@@ -223,8 +220,8 @@ public class TestDocumentWriter extends
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
TEST_VERSION_CURRENT, new MockAnalyzer(random)));
Document doc = new Document();
-
- doc.add(new Field("preanalyzed", new TokenStream() {
+
+ doc.add(new TextField("preanalyzed", new TokenStream() {
private String[] tokens = new String[] {"term1", "term2", "term3", "term2"};
private int index = 0;
@@ -241,7 +238,7 @@ public class TestDocumentWriter extends
}
}
- }, TermVector.NO));
+ }));
writer.addDocument(doc);
writer.commit();
@@ -274,11 +271,15 @@ public class TestDocumentWriter extends
public void testMixedTermVectorSettingsSameField() throws Exception {
Document doc = new Document();
// f1 first without tv then with tv
- doc.add(newField("f1", "v1", Store.YES, Index.NOT_ANALYZED, TermVector.NO));
- doc.add(newField("f1", "v2", Store.YES, Index.NOT_ANALYZED, TermVector.WITH_POSITIONS_OFFSETS));
+ doc.add(newField("f1", "v1", StringField.TYPE_STORED));
+ FieldType customType2 = new FieldType(StringField.TYPE_STORED);
+ customType2.setStoreTermVectors(true);
+ customType2.setStoreTermVectorOffsets(true);
+ customType2.setStoreTermVectorPositions(true);
+ doc.add(newField("f1", "v2", customType2));
// f2 first with tv then without tv
- doc.add(newField("f2", "v1", Store.YES, Index.NOT_ANALYZED, TermVector.WITH_POSITIONS_OFFSETS));
- doc.add(newField("f2", "v2", Store.YES, Index.NOT_ANALYZED, TermVector.NO));
+ doc.add(newField("f2", "v1", customType2));
+ doc.add(newField("f2", "v2", StringField.TYPE_STORED));
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
TEST_VERSION_CURRENT, new MockAnalyzer(random)));
@@ -307,13 +308,18 @@ public class TestDocumentWriter extends
public void testLUCENE_1590() throws Exception {
Document doc = new Document();
// f1 has no norms
- doc.add(newField("f1", "v1", Store.NO, Index.ANALYZED_NO_NORMS));
- doc.add(newField("f1", "v2", Store.YES, Index.NO));
+ FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
+ customType.setOmitNorms(true);
+ FieldType customType2 = new FieldType();
+ customType2.setStored(true);
+ doc.add(newField("f1", "v1", customType));
+ doc.add(newField("f1", "v2", customType2));
// f2 has no TF
- Field f = newField("f2", "v1", Store.NO, Index.ANALYZED);
- f.setIndexOptions(IndexOptions.DOCS_ONLY);
+ FieldType customType3 = new FieldType(TextField.TYPE_UNSTORED);
+ customType3.setIndexOptions(IndexOptions.DOCS_ONLY);
+ Field f = newField("f2", "v1", customType3);
doc.add(f);
- doc.add(newField("f2", "v2", Store.YES, Index.NO));
+ doc.add(newField("f2", "v2", customType2));
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
TEST_VERSION_CURRENT, new MockAnalyzer(random)));
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFieldsReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFieldsReader.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFieldsReader.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFieldsReader.java Mon Aug 29 23:13:10 2011
@@ -24,16 +24,12 @@ import java.util.*;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
+import org.apache.lucene.document.FieldType;
import org.apache.lucene.document.NumericField;
-import org.apache.lucene.document.FieldSelector;
-import org.apache.lucene.document.FieldSelectorResult;
-import org.apache.lucene.document.Fieldable;
-import org.apache.lucene.document.LoadFirstFieldSelector;
-import org.apache.lucene.document.SetBasedFieldSelector;
+import org.apache.lucene.document.StringField;
import org.apache.lucene.index.FieldInfo.IndexOptions;
import org.apache.lucene.index.IndexWriterConfig.OpenMode;
import org.apache.lucene.search.FieldCache;
-import org.apache.lucene.store.AlreadyClosedException;
import org.apache.lucene.store.BufferedIndexInput;
import org.apache.lucene.store.Directory;
import org.apache.lucene.store.IOContext;
@@ -48,7 +44,6 @@ public class TestFieldsReader extends Lu
private static Directory dir;
private static Document testDoc = new Document();
private static FieldInfos fieldInfos = null;
- private final static String TEST_SEGMENT_NAME = "_0";
@BeforeClass
public static void beforeClass() throws Exception {
@@ -61,14 +56,9 @@ public class TestFieldsReader extends Lu
IndexWriter writer = new IndexWriter(dir, conf);
writer.addDocument(testDoc);
writer.close();
- }
-
- @Override
- public void setUp() throws Exception {
- super.setUp();
FaultyIndexInput.doFail = false;
}
-
+
@AfterClass
public static void afterClass() throws Exception {
dir.close();
@@ -76,325 +66,48 @@ public class TestFieldsReader extends Lu
fieldInfos = null;
testDoc = null;
}
+
public void test() throws IOException {
assertTrue(dir != null);
assertTrue(fieldInfos != null);
- FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos);
- assertTrue(reader.size() == 1);
- Document doc = reader.doc(0, null);
+ IndexReader reader = IndexReader.open(dir);
+ Document doc = reader.document(0);
assertTrue(doc != null);
assertTrue(doc.getField(DocHelper.TEXT_FIELD_1_KEY) != null);
- Fieldable field = doc.getField(DocHelper.TEXT_FIELD_2_KEY);
+ Field field = (Field) doc.getField(DocHelper.TEXT_FIELD_2_KEY);
assertTrue(field != null);
- assertTrue(field.isTermVectorStored() == true);
+ assertTrue(field.storeTermVectors() == true);
- assertTrue(field.isStoreOffsetWithTermVector() == true);
- assertTrue(field.isStorePositionWithTermVector() == true);
- assertTrue(field.getOmitNorms() == false);
- assertTrue(field.getIndexOptions() == IndexOptions.DOCS_AND_FREQS_AND_POSITIONS);
+ assertTrue(field.storeTermVectorOffsets() == true);
+ assertTrue(field.storeTermVectorPositions() == true);
+ assertTrue(field.omitNorms() == false);
+ assertTrue(field.indexOptions() == IndexOptions.DOCS_AND_FREQS_AND_POSITIONS);
- field = doc.getField(DocHelper.TEXT_FIELD_3_KEY);
+ field = (Field) doc.getField(DocHelper.TEXT_FIELD_3_KEY);
assertTrue(field != null);
- assertTrue(field.isTermVectorStored() == false);
- assertTrue(field.isStoreOffsetWithTermVector() == false);
- assertTrue(field.isStorePositionWithTermVector() == false);
- assertTrue(field.getOmitNorms() == true);
- assertTrue(field.getIndexOptions() == IndexOptions.DOCS_AND_FREQS_AND_POSITIONS);
+ assertTrue(field.storeTermVectors() == false);
+ assertTrue(field.storeTermVectorOffsets() == false);
+ assertTrue(field.storeTermVectorPositions() == false);
+ assertTrue(field.omitNorms() == true);
+ assertTrue(field.indexOptions() == IndexOptions.DOCS_AND_FREQS_AND_POSITIONS);
- field = doc.getField(DocHelper.NO_TF_KEY);
+ field = (Field) doc.getField(DocHelper.NO_TF_KEY);
assertTrue(field != null);
- assertTrue(field.isTermVectorStored() == false);
- assertTrue(field.isStoreOffsetWithTermVector() == false);
- assertTrue(field.isStorePositionWithTermVector() == false);
- assertTrue(field.getOmitNorms() == false);
- assertTrue(field.getIndexOptions() == IndexOptions.DOCS_ONLY);
- reader.close();
- }
-
-
- public void testLazyFields() throws Exception {
- assertTrue(dir != null);
- assertTrue(fieldInfos != null);
- FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos);
- assertTrue(reader.size() == 1);
- Set<String> loadFieldNames = new HashSet<String>();
- loadFieldNames.add(DocHelper.TEXT_FIELD_1_KEY);
- loadFieldNames.add(DocHelper.TEXT_FIELD_UTF1_KEY);
- Set<String> lazyFieldNames = new HashSet<String>();
- //new String[]{DocHelper.LARGE_LAZY_FIELD_KEY, DocHelper.LAZY_FIELD_KEY, DocHelper.LAZY_FIELD_BINARY_KEY};
- lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY);
- lazyFieldNames.add(DocHelper.LAZY_FIELD_KEY);
- lazyFieldNames.add(DocHelper.LAZY_FIELD_BINARY_KEY);
- lazyFieldNames.add(DocHelper.TEXT_FIELD_UTF2_KEY);
- SetBasedFieldSelector fieldSelector = new SetBasedFieldSelector(loadFieldNames, lazyFieldNames);
- Document doc = reader.doc(0, fieldSelector);
- assertTrue("doc is null and it shouldn't be", doc != null);
- Fieldable field = doc.getFieldable(DocHelper.LAZY_FIELD_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("field is not lazy and it should be", field.isLazy());
- String value = field.stringValue();
- assertTrue("value is null and it shouldn't be", value != null);
- assertTrue(value + " is not equal to " + DocHelper.LAZY_FIELD_TEXT, value.equals(DocHelper.LAZY_FIELD_TEXT) == true);
- assertTrue("calling stringValue() twice should give same reference", field.stringValue() == field.stringValue());
-
- field = doc.getFieldable(DocHelper.TEXT_FIELD_1_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("Field is lazy and it should not be", field.isLazy() == false);
- field = doc.getFieldable(DocHelper.TEXT_FIELD_UTF1_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("Field is lazy and it should not be", field.isLazy() == false);
- assertTrue(field.stringValue() + " is not equal to " + DocHelper.FIELD_UTF1_TEXT, field.stringValue().equals(DocHelper.FIELD_UTF1_TEXT) == true);
-
- field = doc.getFieldable(DocHelper.TEXT_FIELD_UTF2_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("Field is lazy and it should not be", field.isLazy() == true);
- assertTrue(field.stringValue() + " is not equal to " + DocHelper.FIELD_UTF2_TEXT, field.stringValue().equals(DocHelper.FIELD_UTF2_TEXT) == true);
-
- field = doc.getFieldable(DocHelper.LAZY_FIELD_BINARY_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("stringValue isn't null for lazy binary field", field.stringValue() == null);
-
- byte [] bytes = field.getBinaryValue();
- assertTrue("bytes is null and it shouldn't be", bytes != null);
- assertTrue("", DocHelper.LAZY_FIELD_BINARY_BYTES.length == bytes.length);
- assertTrue("calling binaryValue() twice should give same reference", field.getBinaryValue() == field.getBinaryValue());
- for (int i = 0; i < bytes.length; i++) {
- assertTrue("byte[" + i + "] is mismatched", bytes[i] == DocHelper.LAZY_FIELD_BINARY_BYTES[i]);
-
- }
- reader.close();
- }
-
- public void testLatentFields() throws Exception {
- assertTrue(dir != null);
- assertTrue(fieldInfos != null);
- FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos);
- assertTrue(reader.size() == 1);
- Set<String> loadFieldNames = new HashSet<String>();
- loadFieldNames.add(DocHelper.TEXT_FIELD_1_KEY);
- loadFieldNames.add(DocHelper.TEXT_FIELD_UTF1_KEY);
- Set<String> lazyFieldNames = new HashSet<String>();
- //new String[]{DocHelper.LARGE_LAZY_FIELD_KEY, DocHelper.LAZY_FIELD_KEY, DocHelper.LAZY_FIELD_BINARY_KEY};
- lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY);
- lazyFieldNames.add(DocHelper.LAZY_FIELD_KEY);
- lazyFieldNames.add(DocHelper.LAZY_FIELD_BINARY_KEY);
- lazyFieldNames.add(DocHelper.TEXT_FIELD_UTF2_KEY);
-
- // Use LATENT instead of LAZY
- SetBasedFieldSelector fieldSelector = new SetBasedFieldSelector(loadFieldNames, lazyFieldNames) {
- @Override
- public FieldSelectorResult accept(String fieldName) {
- final FieldSelectorResult result = super.accept(fieldName);
- if (result == FieldSelectorResult.LAZY_LOAD) {
- return FieldSelectorResult.LATENT;
- } else {
- return result;
- }
- }
- };
-
- Document doc = reader.doc(0, fieldSelector);
- assertTrue("doc is null and it shouldn't be", doc != null);
- Fieldable field = doc.getFieldable(DocHelper.LAZY_FIELD_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("field is not lazy and it should be", field.isLazy());
- String value = field.stringValue();
- assertTrue("value is null and it shouldn't be", value != null);
- assertTrue(value + " is not equal to " + DocHelper.LAZY_FIELD_TEXT, value.equals(DocHelper.LAZY_FIELD_TEXT) == true);
- assertTrue("calling stringValue() twice should give different references", field.stringValue() != field.stringValue());
-
- field = doc.getFieldable(DocHelper.TEXT_FIELD_1_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("Field is lazy and it should not be", field.isLazy() == false);
- assertTrue("calling stringValue() twice should give same reference", field.stringValue() == field.stringValue());
-
- field = doc.getFieldable(DocHelper.TEXT_FIELD_UTF1_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("Field is lazy and it should not be", field.isLazy() == false);
- assertTrue(field.stringValue() + " is not equal to " + DocHelper.FIELD_UTF1_TEXT, field.stringValue().equals(DocHelper.FIELD_UTF1_TEXT) == true);
- assertTrue("calling stringValue() twice should give same reference", field.stringValue() == field.stringValue());
-
- field = doc.getFieldable(DocHelper.TEXT_FIELD_UTF2_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("Field is lazy and it should not be", field.isLazy() == true);
- assertTrue(field.stringValue() + " is not equal to " + DocHelper.FIELD_UTF2_TEXT, field.stringValue().equals(DocHelper.FIELD_UTF2_TEXT) == true);
- assertTrue("calling stringValue() twice should give different references", field.stringValue() != field.stringValue());
-
- field = doc.getFieldable(DocHelper.LAZY_FIELD_BINARY_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("stringValue isn't null for lazy binary field", field.stringValue() == null);
- assertTrue("calling binaryValue() twice should give different references", field.getBinaryValue() != field.getBinaryValue());
-
- byte [] bytes = field.getBinaryValue();
- assertTrue("bytes is null and it shouldn't be", bytes != null);
- assertTrue("", DocHelper.LAZY_FIELD_BINARY_BYTES.length == bytes.length);
- for (int i = 0; i < bytes.length; i++) {
- assertTrue("byte[" + i + "] is mismatched", bytes[i] == DocHelper.LAZY_FIELD_BINARY_BYTES[i]);
-
- }
- reader.close();
- }
-
-
-
-
- public void testLazyFieldsAfterClose() throws Exception {
- assertTrue(dir != null);
- assertTrue(fieldInfos != null);
- FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos);
- assertTrue(reader.size() == 1);
- Set<String> loadFieldNames = new HashSet<String>();
- loadFieldNames.add(DocHelper.TEXT_FIELD_1_KEY);
- loadFieldNames.add(DocHelper.TEXT_FIELD_UTF1_KEY);
- Set<String> lazyFieldNames = new HashSet<String>();
- lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY);
- lazyFieldNames.add(DocHelper.LAZY_FIELD_KEY);
- lazyFieldNames.add(DocHelper.LAZY_FIELD_BINARY_KEY);
- lazyFieldNames.add(DocHelper.TEXT_FIELD_UTF2_KEY);
- SetBasedFieldSelector fieldSelector = new SetBasedFieldSelector(loadFieldNames, lazyFieldNames);
- Document doc = reader.doc(0, fieldSelector);
- assertTrue("doc is null and it shouldn't be", doc != null);
- Fieldable field = doc.getFieldable(DocHelper.LAZY_FIELD_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("field is not lazy and it should be", field.isLazy());
- reader.close();
- try {
- field.stringValue();
- fail("did not hit AlreadyClosedException as expected");
- } catch (AlreadyClosedException e) {
- // expected
- }
- }
-
- public void testLoadFirst() throws Exception {
- assertTrue(dir != null);
- assertTrue(fieldInfos != null);
- FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos);
- assertTrue(reader.size() == 1);
- LoadFirstFieldSelector fieldSelector = new LoadFirstFieldSelector();
- Document doc = reader.doc(0, fieldSelector);
- assertTrue("doc is null and it shouldn't be", doc != null);
- int count = 0;
- List<Fieldable> l = doc.getFields();
- for (final Fieldable fieldable : l ) {
- Field field = (Field) fieldable;
-
- assertTrue("field is null and it shouldn't be", field != null);
- String sv = field.stringValue();
- assertTrue("sv is null and it shouldn't be", sv != null);
- count++;
- }
- assertTrue(count + " does not equal: " + 1, count == 1);
+ assertTrue(field.storeTermVectors() == false);
+ assertTrue(field.storeTermVectorOffsets() == false);
+ assertTrue(field.storeTermVectorPositions() == false);
+ assertTrue(field.omitNorms() == false);
+ assertTrue(field.indexOptions() == IndexOptions.DOCS_ONLY);
+
+ DocumentStoredFieldVisitor visitor = new DocumentStoredFieldVisitor(DocHelper.TEXT_FIELD_3_KEY);
+ reader.document(0, visitor);
+ final List<IndexableField> fields = visitor.getDocument().getFields();
+ assertEquals(1, fields.size());
+ assertEquals(DocHelper.TEXT_FIELD_3_KEY, fields.get(0).name());
reader.close();
}
- /**
- * Not really a test per se, but we should have some way of assessing whether this is worthwhile.
- * <p/>
- * Must test using a File based directory
- *
- * @throws Exception
- */
- public void testLazyPerformance() throws Exception {
- String userName = System.getProperty("user.name");
- File file = _TestUtil.getTempDir("lazyDir" + userName);
- Directory tmpDir = newFSDirectory(file);
- assertTrue(tmpDir != null);
-
- IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE).setMergePolicy(newLogMergePolicy());
- ((LogMergePolicy) conf.getMergePolicy()).setUseCompoundFile(false);
- IndexWriter writer = new IndexWriter(tmpDir, conf);
- writer.addDocument(testDoc);
- writer.close();
-
- assertTrue(fieldInfos != null);
- FieldsReader reader;
- long lazyTime = 0;
- long regularTime = 0;
- int length = 10;
- Set<String> lazyFieldNames = new HashSet<String>();
- lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY);
- SetBasedFieldSelector fieldSelector = new SetBasedFieldSelector(Collections. <String> emptySet(), lazyFieldNames);
-
- for (int i = 0; i < length; i++) {
- reader = new FieldsReader(tmpDir, TEST_SEGMENT_NAME, fieldInfos);
- assertTrue(reader.size() == 1);
-
- Document doc;
- doc = reader.doc(0, null);//Load all of them
- assertTrue("doc is null and it shouldn't be", doc != null);
- Fieldable field = doc.getFieldable(DocHelper.LARGE_LAZY_FIELD_KEY);
- assertTrue("field is null and it shouldn't be", field != null);
- assertTrue("field is lazy", field.isLazy() == false);
- String value;
- long start;
- long finish;
- start = System.currentTimeMillis();
- //On my machine this was always 0ms.
- value = field.stringValue();
- finish = System.currentTimeMillis();
- assertTrue("value is null and it shouldn't be", value != null);
- regularTime += (finish - start);
- reader.close();
- reader = null;
- doc = null;
- //Hmmm, are we still in cache???
- System.gc();
- reader = new FieldsReader(tmpDir, TEST_SEGMENT_NAME, fieldInfos);
- doc = reader.doc(0, fieldSelector);
- field = doc.getFieldable(DocHelper.LARGE_LAZY_FIELD_KEY);
- assertTrue("field is not lazy", field.isLazy() == true);
- start = System.currentTimeMillis();
- //On my machine this took around 50 - 70ms
- value = field.stringValue();
- finish = System.currentTimeMillis();
- assertTrue("value is null and it shouldn't be", value != null);
- lazyTime += (finish - start);
- reader.close();
-
- }
- tmpDir.close();
- if (VERBOSE) {
- System.out.println("Average Non-lazy time (should be very close to zero): " + regularTime / length + " ms for " + length + " reads");
- System.out.println("Average Lazy Time (should be greater than zero): " + lazyTime / length + " ms for " + length + " reads");
- }
- }
-
- public void testLoadSize() throws IOException {
- FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos);
- Document doc;
-
- doc = reader.doc(0, new FieldSelector(){
- public FieldSelectorResult accept(String fieldName) {
- if (fieldName.equals(DocHelper.TEXT_FIELD_1_KEY) ||
- fieldName.equals(DocHelper.LAZY_FIELD_BINARY_KEY))
- return FieldSelectorResult.SIZE;
- else if (fieldName.equals(DocHelper.TEXT_FIELD_3_KEY))
- return FieldSelectorResult.LOAD;
- else
- return FieldSelectorResult.NO_LOAD;
- }
- });
- Fieldable f1 = doc.getFieldable(DocHelper.TEXT_FIELD_1_KEY);
- Fieldable f3 = doc.getFieldable(DocHelper.TEXT_FIELD_3_KEY);
- Fieldable fb = doc.getFieldable(DocHelper.LAZY_FIELD_BINARY_KEY);
- assertTrue(f1.isBinary());
- assertTrue(!f3.isBinary());
- assertTrue(fb.isBinary());
- assertSizeEquals(2*DocHelper.FIELD_1_TEXT.length(), f1.getBinaryValue());
- assertEquals(DocHelper.FIELD_3_TEXT, f3.stringValue());
- assertSizeEquals(DocHelper.LAZY_FIELD_BINARY_BYTES.length, fb.getBinaryValue());
-
- reader.close();
- }
-
- private void assertSizeEquals(int size, byte[] sizebytes) {
- assertEquals((byte) (size>>>24), sizebytes[0]);
- assertEquals((byte) (size>>>16), sizebytes[1]);
- assertEquals((byte) (size>>> 8), sizebytes[2]);
- assertEquals((byte) size , sizebytes[3]);
- }
public static class FaultyFSDirectory extends Directory {
@@ -528,7 +241,7 @@ public class TestFieldsReader extends Lu
final NumericField.DataType[] typeAnswers = new NumericField.DataType[numDocs];
for(int id=0;id<numDocs;id++) {
Document doc = new Document();
- NumericField nf = new NumericField("nf", Field.Store.YES, false);
+ NumericField nf = new NumericField("nf", NumericField.TYPE_STORED);
doc.add(nf);
final Number answer;
final NumericField.DataType typeAnswer;
@@ -561,7 +274,7 @@ public class TestFieldsReader extends Lu
}
answers[id] = answer;
typeAnswers[id] = typeAnswer;
- doc.add(new NumericField("id", Integer.MAX_VALUE, Field.Store.NO, true).setIntValue(id));
+ doc.add(new NumericField("id", Integer.MAX_VALUE).setIntValue(id));
w.addDocument(doc);
}
final IndexReader r = w.getReader();
@@ -573,15 +286,31 @@ public class TestFieldsReader extends Lu
final int[] ids = FieldCache.DEFAULT.getInts(sub, "id");
for(int docID=0;docID<sub.numDocs();docID++) {
final Document doc = sub.document(docID);
- final Fieldable f = doc.getFieldable("nf");
+ final Field f = (Field) doc.getField("nf");
assertTrue("got f=" + f, f instanceof NumericField);
final NumericField nf = (NumericField) f;
- assertEquals(answers[ids[docID]], nf.getNumericValue());
- assertSame(typeAnswers[ids[docID]], nf.getDataType());
+ assertEquals(answers[ids[docID]], nf.numericValue());
+ assertSame(typeAnswers[ids[docID]], nf.numericDataType());
}
}
r.close();
dir.close();
}
-
+
+ public void testIndexedBit() throws Exception {
+ Directory dir = newDirectory();
+ RandomIndexWriter w = new RandomIndexWriter(random, dir);
+ Document doc = new Document();
+ FieldType onlyStored = new FieldType();
+ onlyStored.setStored(true);
+ doc.add(new Field("field", onlyStored, "value"));
+ doc.add(new Field("field2", StringField.TYPE_STORED, "value"));
+ w.addDocument(doc);
+ IndexReader r = w.getReader();
+ w.close();
+ assertFalse(r.document(0).getField("field").indexed());
+ assertTrue(r.document(0).getField("field2").indexed());
+ r.close();
+ dir.close();
+ }
}
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFilterIndexReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFilterIndexReader.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFilterIndexReader.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFilterIndexReader.java Mon Aug 29 23:13:10 2011
@@ -23,7 +23,7 @@ import org.apache.lucene.util.LuceneTest
import org.apache.lucene.store.Directory;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.util.BytesRef;
import org.apache.lucene.util.Bits;
@@ -129,15 +129,15 @@ public class TestFilterIndexReader exten
IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
Document d1 = new Document();
- d1.add(newField("default","one two", Field.Store.YES, Field.Index.ANALYZED));
+ d1.add(newField("default","one two", TextField.TYPE_STORED));
writer.addDocument(d1);
Document d2 = new Document();
- d2.add(newField("default","one three", Field.Store.YES, Field.Index.ANALYZED));
+ d2.add(newField("default","one three", TextField.TYPE_STORED));
writer.addDocument(d2);
Document d3 = new Document();
- d3.add(newField("default","two four", Field.Store.YES, Field.Index.ANALYZED));
+ d3.add(newField("default","two four", TextField.TYPE_STORED));
writer.addDocument(d3);
writer.close();
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFlex.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFlex.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFlex.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestFlex.java Mon Aug 29 23:13:10 2011
@@ -39,10 +39,10 @@ public class TestFlex extends LuceneTest
for(int iter=0;iter<2;iter++) {
if (iter == 0) {
Document doc = new Document();
- doc.add(newField("field1", "this is field1", Field.Store.NO, Field.Index.ANALYZED));
- doc.add(newField("field2", "this is field2", Field.Store.NO, Field.Index.ANALYZED));
- doc.add(newField("field3", "aaa", Field.Store.NO, Field.Index.ANALYZED));
- doc.add(newField("field4", "bbb", Field.Store.NO, Field.Index.ANALYZED));
+ doc.add(newField("field1", "this is field1", TextField.TYPE_UNSTORED));
+ doc.add(newField("field2", "this is field2", TextField.TYPE_UNSTORED));
+ doc.add(newField("field3", "aaa", TextField.TYPE_UNSTORED));
+ doc.add(newField("field4", "bbb", TextField.TYPE_UNSTORED));
for(int i=0;i<DOC_COUNT;i++) {
w.addDocument(doc);
}
@@ -66,7 +66,7 @@ public class TestFlex extends LuceneTest
IndexWriter w = new IndexWriter(d, newIndexWriterConfig(TEST_VERSION_CURRENT,
new MockAnalyzer(random)).setCodecProvider(_TestUtil.alwaysCodec("Standard")));
Document doc = new Document();
- doc.add(newField("f", "a b c", Field.Store.NO, Field.Index.ANALYZED));
+ doc.add(newField("f", "a b c", TextField.TYPE_UNSTORED));
w.addDocument(doc);
IndexReader r = w.getReader();
TermsEnum terms = r.getSequentialSubReaders()[0].fields().terms("f").iterator();
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestGlobalFieldNumbers.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestGlobalFieldNumbers.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestGlobalFieldNumbers.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestGlobalFieldNumbers.java Mon Aug 29 23:13:10 2011
@@ -27,11 +27,11 @@ import java.util.SortedMap;
import java.util.TreeMap;
import org.apache.lucene.analysis.MockAnalyzer;
+import org.apache.lucene.document.BinaryField;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
-import org.apache.lucene.document.Field.Index;
-import org.apache.lucene.document.Field.Store;
-import org.apache.lucene.document.Field.TermVector;
+import org.apache.lucene.document.FieldType;
+import org.apache.lucene.document.TextField;
import org.apache.lucene.index.FieldInfos.FieldNumberBiMap;
import org.apache.lucene.index.codecs.DefaultSegmentInfosWriter;
import org.apache.lucene.store.Directory;
@@ -50,10 +50,8 @@ public class TestGlobalFieldNumbers exte
new MockAnalyzer(random));
IndexWriter writer = new IndexWriter(dir, config);
Document d = new Document();
- d.add(new Field("f1", "d1 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d1 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d1 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d1 second field"));
writer.addDocument(d);
for (String string : writer.getIndexFileNames()) {
assertFalse(string.endsWith(".fnx"));
@@ -67,9 +65,8 @@ public class TestGlobalFieldNumbers exte
assertFNXFiles(dir, "1.fnx");
d = new Document();
- d.add(new Field("f1", "d2 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d2 first field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3 }));
writer.addDocument(d);
writer.commit();
files = writer.getIndexFileNames();
@@ -86,11 +83,9 @@ public class TestGlobalFieldNumbers exte
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
TEST_VERSION_CURRENT, new MockAnalyzer(random)));
Document d = new Document();
- d.add(new Field("f1", "d3 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d3 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3, 4, 5 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d3 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d3 second field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3, 4, 5 }));
writer.addDocument(d);
writer.close();
Collection<String> files = writer.getIndexFileNames();
@@ -122,17 +117,14 @@ public class TestGlobalFieldNumbers exte
new MockAnalyzer(random));
IndexWriter writer = new IndexWriter(dir, config);
Document d = new Document();
- d.add(new Field("f1", "d1 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d1 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d1 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d1 second field"));
writer.addDocument(d);
writer.commit();
assertFNXFiles(dir, "1.fnx");
d = new Document();
- d.add(new Field("f1", "d2 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d2 first field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3 }));
writer.addDocument(d);
writer.commit();
assertFNXFiles(dir, "2.fnx");
@@ -166,17 +158,14 @@ public class TestGlobalFieldNumbers exte
TEST_VERSION_CURRENT, new MockAnalyzer(random)).setIndexDeletionPolicy(
new KeepAllDeletionPolicy()));
Document d = new Document();
- d.add(new Field("f1", "d1 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d1 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d1 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d1 second field"));
writer.addDocument(d);
writer.commit();
assertFNXFiles(dir, "1.fnx");
d = new Document();
- d.add(new Field("f1", "d2 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d2 first field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3 }));
writer.addDocument(d);
writer.commit();
writer.commit();
@@ -190,11 +179,9 @@ public class TestGlobalFieldNumbers exte
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
TEST_VERSION_CURRENT, new MockAnalyzer(random)));
Document d = new Document();
- d.add(new Field("f1", "d3 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d3 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3, 4, 5 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d3 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d3 second field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3, 4, 5 }));
writer.addDocument(d);
writer.close();
assertFNXFiles(dir, "2.fnx");
@@ -217,17 +204,14 @@ public class TestGlobalFieldNumbers exte
TEST_VERSION_CURRENT, new MockAnalyzer(random)).setIndexDeletionPolicy(
new KeepAllDeletionPolicy()));
Document d = new Document();
- d.add(new Field("f1", "d1 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f2", "d1 second field", Store.YES, Index.ANALYZED,
- TermVector.NO));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d1 first field"));
+ d.add(new Field("f2", TextField.TYPE_STORED, "d1 second field"));
writer.addDocument(d);
writer.commit();
assertFNXFiles(dir, "1.fnx");
d = new Document();
- d.add(new Field("f1", "d2 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d2 first field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3 }));
writer.addDocument(d);
assertFNXFiles(dir, "1.fnx");
writer.close();
@@ -240,9 +224,8 @@ public class TestGlobalFieldNumbers exte
new KeepAllDeletionPolicy()).setIndexCommit(listCommits.get(0)));
d = new Document();
- d.add(new Field("f1", "d2 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
- d.add(new Field("f3", new byte[] { 1, 2, 3 }));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d2 first field"));
+ d.add(new BinaryField("f3", new byte[] { 1, 2, 3 }));
writer.addDocument(d);
writer.commit();
// now we have 3 files since f3 is not present in the first commit
@@ -271,9 +254,13 @@ public class TestGlobalFieldNumbers exte
Document doc = new Document();
final int numFields = 1 + random.nextInt(fieldNames.length);
for (int j = 0; j < numFields; j++) {
+ FieldType customType = new FieldType();
+ customType.setIndexed(true);
+ customType.setTokenized(random.nextBoolean());
+ customType.setOmitNorms(random.nextBoolean());
doc.add(newField(fieldNames[random.nextInt(fieldNames.length)],
_TestUtil.randomRealisticUnicodeString(random),
- Index.toIndex(true, random.nextBoolean(), random.nextBoolean())));
+ customType));
}
writer.addDocument(doc);
@@ -322,9 +309,13 @@ public class TestGlobalFieldNumbers exte
TEST_VERSION_CURRENT, new MockAnalyzer(random)));
Document doc = new Document();
for (String string : fieldNames) {
+ FieldType customType = new FieldType();
+ customType.setIndexed(true);
+ customType.setTokenized(random.nextBoolean());
+ customType.setOmitNorms(random.nextBoolean());
doc.add(newField(string,
_TestUtil.randomRealisticUnicodeString(random),
- Index.toIndex(true, random.nextBoolean(), random.nextBoolean())));
+ customType));
}
writer.addDocument(doc);
@@ -419,8 +410,12 @@ public class TestGlobalFieldNumbers exte
String name = copySortedMap.get(nextField);
assertNotNull(name);
+ FieldType customType = new FieldType();
+ customType.setIndexed(true);
+ customType.setTokenized(random.nextBoolean());
+ customType.setOmitNorms(random.nextBoolean());
doc.add(newField(name, _TestUtil.randomRealisticUnicodeString(random),
- Index.toIndex(true, random.nextBoolean(), random.nextBoolean())));
+ customType));
writer.addDocument(doc);
if (random.nextInt(10) == 0) {
writer.commit();
@@ -480,8 +475,7 @@ public class TestGlobalFieldNumbers exte
}
Document d = new Document();
- d.add(new Field("f1", "d1 first field", Store.YES, Index.ANALYZED,
- TermVector.NO));
+ d.add(new Field("f1", TextField.TYPE_STORED, "d1 first field"));
writer.addDocument(d);
writer.prepareCommit();
// the fnx file should still be under control of the SIS
Modified: lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java?rev=1163047&r1=1163046&r2=1163047&view=diff
==============================================================================
--- lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java (original)
+++ lucene/dev/branches/flexscoring/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java Mon Aug 29 23:13:10 2011
@@ -17,20 +17,21 @@ package org.apache.lucene.index;
* limitations under the License.
*/
-import org.apache.lucene.util.LuceneTestCase;
+import java.io.*;
+import java.util.*;
+
+import org.apache.lucene.analysis.MockAnalyzer;
+import org.apache.lucene.document.Document;
+import org.apache.lucene.document.StringField;
+import org.apache.lucene.document.TextField;
+import org.apache.lucene.index.IndexWriterConfig.OpenMode;
import org.apache.lucene.search.similarities.DefaultSimilarity;
import org.apache.lucene.store.CompoundFileDirectory;
import org.apache.lucene.store.Directory;
import org.apache.lucene.store.IndexInput;
import org.apache.lucene.store.IndexOutput;
import org.apache.lucene.store.MockDirectoryWrapper;
-import org.apache.lucene.analysis.MockAnalyzer;
-import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
-import org.apache.lucene.index.IndexWriterConfig.OpenMode;
-
-import java.io.*;
-import java.util.*;
+import org.apache.lucene.util.LuceneTestCase;
/*
Verify we can read the pre-2.1 file format, do searches
@@ -229,8 +230,8 @@ public class TestIndexFileDeleter extend
private void addDoc(IndexWriter writer, int id) throws IOException
{
Document doc = new Document();
- doc.add(newField("content", "aaa", Field.Store.NO, Field.Index.ANALYZED));
- doc.add(newField("id", Integer.toString(id), Field.Store.YES, Field.Index.NOT_ANALYZED));
+ doc.add(newField("content", "aaa", TextField.TYPE_UNSTORED));
+ doc.add(newField("id", Integer.toString(id), StringField.TYPE_UNSTORED));
writer.addDocument(doc);
}
}