You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by dw...@apache.org on 2012/04/15 16:42:01 UTC
svn commit: r1326351 [3/22] - in /lucene/dev/trunk: ./ dev-tools/eclipse/
lucene/
lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/
lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/custom/
lucene/contrib/high...
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/lucene40/values/TestDocValues.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/lucene40/values/TestDocValues.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/lucene40/values/TestDocValues.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/lucene40/values/TestDocValues.java Sun Apr 15 14:41:44 2012
@@ -20,6 +20,7 @@ package org.apache.lucene.codecs.lucene4
import java.io.IOException;
import java.io.Reader;
import java.util.Comparator;
+import java.util.Random;
import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.TokenStream;
@@ -69,18 +70,18 @@ public class TestDocValues extends Lucen
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Bytes.getWriter(dir, "test", mode, fixedSize, COMP, trackBytes, newIOContext(random),
- random.nextBoolean());
+ DocValuesConsumer w = Bytes.getWriter(dir, "test", mode, fixedSize, COMP, trackBytes, newIOContext(random()),
+ random().nextBoolean());
int maxDoc = 220;
final String[] values = new String[maxDoc];
final int fixedLength = 1 + atLeast(50);
for (int i = 0; i < 100; i++) {
final String s;
- if (i > 0 && random.nextInt(5) <= 2) {
+ if (i > 0 && random().nextInt(5) <= 2) {
// use prior value
- s = values[2 * random.nextInt(i)];
+ s = values[2 * random().nextInt(i)];
} else {
- s = _TestUtil.randomFixedByteLengthUnicodeString(random, fixedSize? fixedLength : 1 + random.nextInt(39));
+ s = _TestUtil.randomFixedByteLengthUnicodeString(random(), fixedSize? fixedLength : 1 + random().nextInt(39));
}
values[2 * i] = s;
@@ -91,7 +92,7 @@ public class TestDocValues extends Lucen
w.finish(maxDoc);
assertEquals(0, trackBytes.get());
- DocValues r = Bytes.getValues(dir, "test", mode, fixedSize, maxDoc, COMP, newIOContext(random));
+ DocValues r = Bytes.getValues(dir, "test", mode, fixedSize, maxDoc, COMP, newIOContext(random()));
// Verify we can load source twice:
for (int iter = 0; iter < 2; iter++) {
@@ -123,6 +124,7 @@ public class TestDocValues extends Lucen
// Lookup random strings:
if (mode == Bytes.Mode.SORTED) {
final int valueCount = ss.getValueCount();
+ Random random = random();
for (int i = 0; i < 1000; i++) {
BytesRef bytesValue = new BytesRef(_TestUtil.randomFixedByteLengthUnicodeString(random, fixedSize? fixedLength : 1 + random.nextInt(39)));
int ord = ss.getOrdByValue(bytesValue, new BytesRef());
@@ -178,14 +180,14 @@ public class TestDocValues extends Lucen
for (int i = 0; i < minMax.length; i++) {
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.VAR_INTS, newIOContext(random));
+ DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.VAR_INTS, newIOContext(random()));
valueHolder.numberValue = minMax[i][0];
w.add(0, valueHolder);
valueHolder.numberValue = minMax[i][1];
w.add(1, valueHolder);
w.finish(2);
assertEquals(0, trackBytes.get());
- DocValues r = Ints.getValues(dir, "test", 2, Type.VAR_INTS, newIOContext(random));
+ DocValues r = Ints.getValues(dir, "test", 2, Type.VAR_INTS, newIOContext(random()));
Source source = getSource(r);
assertEquals(i + " with min: " + minMax[i][0] + " max: " + minMax[i][1],
expectedTypes[i], source.getType());
@@ -214,13 +216,13 @@ public class TestDocValues extends Lucen
byte[] sourceArray = new byte[] {1,2,3};
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_8, newIOContext(random));
+ DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_8, newIOContext(random()));
for (int i = 0; i < sourceArray.length; i++) {
valueHolder.numberValue = (long) sourceArray[i];
w.add(i, valueHolder);
}
w.finish(sourceArray.length);
- DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_8, newIOContext(random));
+ DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_8, newIOContext(random()));
Source source = r.getSource();
assertTrue(source.hasArray());
byte[] loaded = ((byte[])source.getArray());
@@ -237,13 +239,13 @@ public class TestDocValues extends Lucen
short[] sourceArray = new short[] {1,2,3};
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_16, newIOContext(random));
+ DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_16, newIOContext(random()));
for (int i = 0; i < sourceArray.length; i++) {
valueHolder.numberValue = (long) sourceArray[i];
w.add(i, valueHolder);
}
w.finish(sourceArray.length);
- DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_16, newIOContext(random));
+ DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_16, newIOContext(random()));
Source source = r.getSource();
assertTrue(source.hasArray());
short[] loaded = ((short[])source.getArray());
@@ -260,13 +262,13 @@ public class TestDocValues extends Lucen
long[] sourceArray = new long[] {1,2,3};
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_64, newIOContext(random));
+ DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_64, newIOContext(random()));
for (int i = 0; i < sourceArray.length; i++) {
valueHolder.numberValue = sourceArray[i];
w.add(i, valueHolder);
}
w.finish(sourceArray.length);
- DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_64, newIOContext(random));
+ DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_64, newIOContext(random()));
Source source = r.getSource();
assertTrue(source.hasArray());
long[] loaded = ((long[])source.getArray());
@@ -283,13 +285,13 @@ public class TestDocValues extends Lucen
int[] sourceArray = new int[] {1,2,3};
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_32, newIOContext(random));
+ DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, Type.FIXED_INTS_32, newIOContext(random()));
for (int i = 0; i < sourceArray.length; i++) {
valueHolder.numberValue = (long) sourceArray[i];
w.add(i, valueHolder);
}
w.finish(sourceArray.length);
- DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_32, newIOContext(random));
+ DocValues r = Ints.getValues(dir, "test", sourceArray.length, Type.FIXED_INTS_32, newIOContext(random()));
Source source = r.getSource();
assertTrue(source.hasArray());
int[] loaded = ((int[])source.getArray());
@@ -306,13 +308,13 @@ public class TestDocValues extends Lucen
float[] sourceArray = new float[] {1,2,3};
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Floats.getWriter(dir, "test", trackBytes, newIOContext(random), Type.FLOAT_32);
+ DocValuesConsumer w = Floats.getWriter(dir, "test", trackBytes, newIOContext(random()), Type.FLOAT_32);
for (int i = 0; i < sourceArray.length; i++) {
valueHolder.numberValue = sourceArray[i];
w.add(i, valueHolder);
}
w.finish(sourceArray.length);
- DocValues r = Floats.getValues(dir, "test", 3, newIOContext(random), Type.FLOAT_32);
+ DocValues r = Floats.getValues(dir, "test", 3, newIOContext(random()), Type.FLOAT_32);
Source source = r.getSource();
assertTrue(source.hasArray());
float[] loaded = ((float[])source.getArray());
@@ -329,13 +331,13 @@ public class TestDocValues extends Lucen
double[] sourceArray = new double[] {1,2,3};
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Floats.getWriter(dir, "test", trackBytes, newIOContext(random), Type.FLOAT_64);
+ DocValuesConsumer w = Floats.getWriter(dir, "test", trackBytes, newIOContext(random()), Type.FLOAT_64);
for (int i = 0; i < sourceArray.length; i++) {
valueHolder.numberValue = sourceArray[i];
w.add(i, valueHolder);
}
w.finish(sourceArray.length);
- DocValues r = Floats.getValues(dir, "test", 3, newIOContext(random), Type.FLOAT_64);
+ DocValues r = Floats.getValues(dir, "test", 3, newIOContext(random()), Type.FLOAT_64);
Source source = r.getSource();
assertTrue(source.hasArray());
double[] loaded = ((double[])source.getArray());
@@ -350,22 +352,22 @@ public class TestDocValues extends Lucen
private void testInts(Type type, int maxBit) throws IOException {
DocValueHolder valueHolder = new DocValueHolder();
long maxV = 1;
- final int NUM_VALUES = 333 + random.nextInt(333);
+ final int NUM_VALUES = 333 + random().nextInt(333);
final long[] values = new long[NUM_VALUES];
for (int rx = 1; rx < maxBit; rx++, maxV *= 2) {
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, type, newIOContext(random));
+ DocValuesConsumer w = Ints.getWriter(dir, "test", trackBytes, type, newIOContext(random()));
for (int i = 0; i < NUM_VALUES; i++) {
- final long v = random.nextLong() % (1 + maxV);
+ final long v = random().nextLong() % (1 + maxV);
valueHolder.numberValue = values[i] = v;
w.add(i, valueHolder);
}
- final int additionalDocs = 1 + random.nextInt(9);
+ final int additionalDocs = 1 + random().nextInt(9);
w.finish(NUM_VALUES + additionalDocs);
assertEquals(0, trackBytes.get());
- DocValues r = Ints.getValues(dir, "test", NUM_VALUES + additionalDocs, type, newIOContext(random));
+ DocValues r = Ints.getValues(dir, "test", NUM_VALUES + additionalDocs, type, newIOContext(random()));
for (int iter = 0; iter < 2; iter++) {
Source s = getSource(r);
assertEquals(type, s.getType());
@@ -388,20 +390,20 @@ public class TestDocValues extends Lucen
DocValueHolder valueHolder = new DocValueHolder();
Directory dir = newDirectory();
final Counter trackBytes = Counter.newCounter();
- DocValuesConsumer w = Floats.getWriter(dir, "test", trackBytes, newIOContext(random), type);
- final int NUM_VALUES = 777 + random.nextInt(777);
+ DocValuesConsumer w = Floats.getWriter(dir, "test", trackBytes, newIOContext(random()), type);
+ final int NUM_VALUES = 777 + random().nextInt(777);
final double[] values = new double[NUM_VALUES];
for (int i = 0; i < NUM_VALUES; i++) {
- final double v = type == Type.FLOAT_32 ? random.nextFloat() : random
+ final double v = type == Type.FLOAT_32 ? random().nextFloat() : random()
.nextDouble();
valueHolder.numberValue = values[i] = v;
w.add(i, valueHolder);
}
- final int additionalValues = 1 + random.nextInt(10);
+ final int additionalValues = 1 + random().nextInt(10);
w.finish(NUM_VALUES + additionalValues);
assertEquals(0, trackBytes.get());
- DocValues r = Floats.getValues(dir, "test", NUM_VALUES + additionalValues, newIOContext(random), type);
+ DocValues r = Floats.getValues(dir, "test", NUM_VALUES + additionalValues, newIOContext(random()), type);
for (int iter = 0; iter < 2; iter++) {
Source s = getSource(r);
for (int i = 0; i < NUM_VALUES; i++) {
@@ -419,7 +421,7 @@ public class TestDocValues extends Lucen
private Source getSource(DocValues values) throws IOException {
// getSource uses cache internally
- switch(random.nextInt(5)) {
+ switch(random().nextInt(5)) {
case 3:
return values.load();
case 2:
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/perfield/TestPerFieldPostingsFormat.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/perfield/TestPerFieldPostingsFormat.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/perfield/TestPerFieldPostingsFormat.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/perfield/TestPerFieldPostingsFormat.java Sun Apr 15 14:41:44 2012
@@ -96,7 +96,7 @@ public class TestPerFieldPostingsFormat
public void testMergeUnusedPerFieldCodec() throws IOException {
Directory dir = newDirectory();
IndexWriterConfig iwconf = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE).setCodec(new MockCodec());
+ new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE).setCodec(new MockCodec());
IndexWriter writer = newWriter(dir, iwconf);
addDocs(writer, 10);
writer.commit();
@@ -123,7 +123,7 @@ public class TestPerFieldPostingsFormat
System.out.println("TEST: make new index");
}
IndexWriterConfig iwconf = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE).setCodec(new MockCodec());
+ new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE).setCodec(new MockCodec());
iwconf.setMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH);
//((LogMergePolicy) iwconf.getMergePolicy()).setMergeFactor(10);
IndexWriter writer = newWriter(dir, iwconf);
@@ -142,7 +142,7 @@ public class TestPerFieldPostingsFormat
assertQuery(new Term("content", "aaa"), dir, 10);
Lucene40Codec codec = (Lucene40Codec)iwconf.getCodec();
- iwconf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ iwconf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setOpenMode(OpenMode.APPEND).setCodec(codec);
//((LogMergePolicy) iwconf.getMergePolicy()).setUseCompoundFile(false);
//((LogMergePolicy) iwconf.getMergePolicy()).setMergeFactor(10);
@@ -234,28 +234,28 @@ public class TestPerFieldPostingsFormat
*/
@Test
public void testStressPerFieldCodec() throws IOException {
- Directory dir = newDirectory(random);
+ Directory dir = newDirectory(random());
final int docsPerRound = 97;
int numRounds = atLeast(1);
for (int i = 0; i < numRounds; i++) {
- int num = _TestUtil.nextInt(random, 30, 60);
- IndexWriterConfig config = newIndexWriterConfig(random,
- TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ int num = _TestUtil.nextInt(random(), 30, 60);
+ IndexWriterConfig config = newIndexWriterConfig(random(),
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()));
config.setOpenMode(OpenMode.CREATE_OR_APPEND);
IndexWriter writer = newWriter(dir, config);
for (int j = 0; j < docsPerRound; j++) {
final Document doc = new Document();
for (int k = 0; k < num; k++) {
FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
- customType.setTokenized(random.nextBoolean());
- customType.setOmitNorms(random.nextBoolean());
+ customType.setTokenized(random().nextBoolean());
+ customType.setOmitNorms(random().nextBoolean());
Field field = newField("" + k, _TestUtil
- .randomRealisticUnicodeString(random, 128), customType);
+ .randomRealisticUnicodeString(random(), 128), customType);
doc.add(field);
}
writer.addDocument(doc);
}
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
writer.forceMerge(1);
}
writer.commit();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/Test10KPulsings.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/Test10KPulsings.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/Test10KPulsings.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/Test10KPulsings.java Sun Apr 15 14:41:44 2012
@@ -56,13 +56,13 @@ public class Test10KPulsings extends Luc
File f = _TestUtil.getTempDir("10kpulsed");
MockDirectoryWrapper dir = newFSDirectory(f);
dir.setCheckIndexOnClose(false); // we do this ourselves explicitly
- RandomIndexWriter iw = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setCodec(cp));
+ RandomIndexWriter iw = new RandomIndexWriter(random(), dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setCodec(cp));
Document document = new Document();
FieldType ft = new FieldType(TextField.TYPE_STORED);
- switch(_TestUtil.nextInt(random, 0, 2)) {
+ switch(_TestUtil.nextInt(random(), 0, 2)) {
case 0: ft.setIndexOptions(IndexOptions.DOCS_ONLY); break;
case 1: ft.setIndexOptions(IndexOptions.DOCS_AND_FREQS); break;
default: ft.setIndexOptions(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS); break;
@@ -87,7 +87,7 @@ public class Test10KPulsings extends Luc
for (int i = 0; i < 10050; i++) {
String expected = df.format(i);
assertEquals(expected, te.next().utf8ToString());
- de = _TestUtil.docs(random, te, null, de, false);
+ de = _TestUtil.docs(random(), te, null, de, false);
assertTrue(de.nextDoc() != DocIdSetIterator.NO_MORE_DOCS);
assertEquals(DocIdSetIterator.NO_MORE_DOCS, de.nextDoc());
}
@@ -101,19 +101,19 @@ public class Test10KPulsings extends Luc
*/
public void test10kNotPulsed() throws Exception {
// we always run this test with pulsing codec.
- int freqCutoff = _TestUtil.nextInt(random, 1, 10);
+ int freqCutoff = _TestUtil.nextInt(random(), 1, 10);
Codec cp = _TestUtil.alwaysPostingsFormat(new Pulsing40PostingsFormat(freqCutoff));
File f = _TestUtil.getTempDir("10knotpulsed");
MockDirectoryWrapper dir = newFSDirectory(f);
dir.setCheckIndexOnClose(false); // we do this ourselves explicitly
- RandomIndexWriter iw = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setCodec(cp));
+ RandomIndexWriter iw = new RandomIndexWriter(random(), dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setCodec(cp));
Document document = new Document();
FieldType ft = new FieldType(TextField.TYPE_STORED);
- switch(_TestUtil.nextInt(random, 0, 2)) {
+ switch(_TestUtil.nextInt(random(), 0, 2)) {
case 0: ft.setIndexOptions(IndexOptions.DOCS_ONLY); break;
case 1: ft.setIndexOptions(IndexOptions.DOCS_AND_FREQS); break;
default: ft.setIndexOptions(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS); break;
@@ -145,7 +145,7 @@ public class Test10KPulsings extends Luc
for (int i = 0; i < 10050; i++) {
String expected = df.format(i);
assertEquals(expected, te.next().utf8ToString());
- de = _TestUtil.docs(random, te, null, de, false);
+ de = _TestUtil.docs(random(), te, null, de, false);
assertTrue(de.nextDoc() != DocIdSetIterator.NO_MORE_DOCS);
assertEquals(DocIdSetIterator.NO_MORE_DOCS, de.nextDoc());
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/TestPulsingReuse.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/TestPulsingReuse.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/TestPulsingReuse.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/codecs/pulsing/TestPulsingReuse.java Sun Apr 15 14:41:44 2012
@@ -48,8 +48,8 @@ public class TestPulsingReuse extends Lu
// we always run this test with pulsing codec.
Codec cp = _TestUtil.alwaysPostingsFormat(new Pulsing40PostingsFormat(1));
Directory dir = newDirectory();
- RandomIndexWriter iw = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setCodec(cp));
+ RandomIndexWriter iw = new RandomIndexWriter(random(), dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setCodec(cp));
Document doc = new Document();
doc.add(new Field("foo", "a b b c c c d e f g g h i i j j k", TextField.TYPE_UNSTORED));
iw.addDocument(doc);
@@ -87,8 +87,8 @@ public class TestPulsingReuse extends Lu
Codec cp = _TestUtil.alwaysPostingsFormat(new NestedPulsingPostingsFormat());
MockDirectoryWrapper dir = newDirectory();
dir.setCheckIndexOnClose(false); // will do this ourselves, custom codec
- RandomIndexWriter iw = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setCodec(cp));
+ RandomIndexWriter iw = new RandomIndexWriter(random(), dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setCodec(cp));
Document doc = new Document();
doc.add(new Field("foo", "a b b c c c d e f g g g h i i j j k l l m m m", TextField.TYPE_UNSTORED));
// note: the reuse is imperfect, here we would have 4 enums (lost reuse when we get an enum for 'm')
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestBinaryDocument.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestBinaryDocument.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestBinaryDocument.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestBinaryDocument.java Sun Apr 15 14:41:44 2012
@@ -51,7 +51,7 @@ public class TestBinaryDocument extends
/** add the doc to a ram index */
Directory dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir);
writer.addDocument(doc);
/** open a reader and fetch the document */
@@ -85,7 +85,7 @@ public class TestBinaryDocument extends
/** add the doc to a ram index */
Directory dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir);
writer.addDocument(doc);
/** open a reader and fetch the document */
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestDocument.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestDocument.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestDocument.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/document/TestDocument.java Sun Apr 15 14:41:44 2012
@@ -171,7 +171,7 @@ public class TestDocument extends Lucene
*/
public void testGetValuesForIndexedDocument() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir);
writer.addDocument(makeDocumentWithFields());
IndexReader reader = writer.getReader();
@@ -256,7 +256,7 @@ public class TestDocument extends Lucene
doc.add(new Field("keyword", "test", StringField.TYPE_STORED));
Directory dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir);
writer.addDocument(doc);
field.setStringValue("id2");
writer.addDocument(doc);
@@ -299,7 +299,7 @@ public class TestDocument extends Lucene
// LUCENE-3682
public void testTransitionAPI() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, dir);
+ RandomIndexWriter w = new RandomIndexWriter(random(), dir);
Document doc = new Document();
doc.add(new Field("stored", "abc", Field.Store.YES, Field.Index.NO));
@@ -363,7 +363,7 @@ public class TestDocument extends Lucene
public void testBoost() throws Exception {
Directory dir = newDirectory();
- IndexWriterConfig iwc = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig iwc = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
iwc.setMergePolicy(newLogMergePolicy());
IndexWriter iw = new IndexWriter(dir, iwc);
Document doc = new Document();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BPostings.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BPostings.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BPostings.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BPostings.java Sun Apr 15 14:41:44 2012
@@ -46,7 +46,7 @@ public class Test2BPostings extends Luce
dir.setCheckIndexOnClose(false); // don't double-checkindex
IndexWriter w = new IndexWriter(dir,
- new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.setRAMBufferSizeMB(256.0)
.setMergeScheduler(new ConcurrentMergeScheduler())
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BTerms.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BTerms.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BTerms.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/Test2BTerms.java Sun Apr 15 14:41:44 2012
@@ -149,7 +149,7 @@ public class Test2BTerms extends LuceneT
System.out.println("Starting Test2B");
final long TERM_COUNT = ((long) Integer.MAX_VALUE) + 100000000;
- final int TERMS_PER_DOC = _TestUtil.nextInt(random, 100000, 1000000);
+ final int TERMS_PER_DOC = _TestUtil.nextInt(random(), 100000, 1000000);
List<BytesRef> savedTerms = null;
@@ -161,7 +161,7 @@ public class Test2BTerms extends LuceneT
if (true) {
IndexWriter w = new IndexWriter(dir,
- new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.setRAMBufferSizeMB(256.0)
.setMergeScheduler(new ConcurrentMergeScheduler())
@@ -175,7 +175,7 @@ public class Test2BTerms extends LuceneT
}
Document doc = new Document();
- final MyTokenStream ts = new MyTokenStream(random, TERMS_PER_DOC);
+ final MyTokenStream ts = new MyTokenStream(random(), TERMS_PER_DOC);
FieldType customType = new FieldType(TextField.TYPE_UNSTORED);
customType.setIndexOptions(IndexOptions.DOCS_ONLY);
@@ -227,13 +227,13 @@ public class Test2BTerms extends LuceneT
System.out.println("TEST: findTerms");
final TermsEnum termsEnum = MultiFields.getTerms(r, "field").iterator(null);
final List<BytesRef> savedTerms = new ArrayList<BytesRef>();
- int nextSave = _TestUtil.nextInt(random, 500000, 1000000);
+ int nextSave = _TestUtil.nextInt(random(), 500000, 1000000);
BytesRef term;
while((term = termsEnum.next()) != null) {
if (--nextSave == 0) {
savedTerms.add(BytesRef.deepCopyOf(term));
System.out.println("TEST: add " + term);
- nextSave = _TestUtil.nextInt(random, 500000, 1000000);
+ nextSave = _TestUtil.nextInt(random(), 500000, 1000000);
}
}
return savedTerms;
@@ -246,7 +246,7 @@ public class Test2BTerms extends LuceneT
TermsEnum termsEnum = MultiFields.getTerms(r, "field").iterator(null);
boolean failed = false;
for(int iter=0;iter<10*terms.size();iter++) {
- final BytesRef term = terms.get(random.nextInt(terms.size()));
+ final BytesRef term = terms.get(random().nextInt(terms.size()));
System.out.println("TEST: search " + term);
final long t0 = System.currentTimeMillis();
final int count = s.search(new TermQuery(new Term("field", term)), 1).totalHits;
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAddIndexes.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAddIndexes.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAddIndexes.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAddIndexes.java Sun Apr 15 14:41:44 2012
@@ -70,7 +70,7 @@ public class TestAddIndexes extends Luce
IndexWriter writer = null;
writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random))
+ new MockAnalyzer(random()))
.setOpenMode(OpenMode.CREATE));
// add 100 documents
addDocs(writer, 100);
@@ -80,7 +80,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
aux,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setMergePolicy(newLogMergePolicy(false))
);
@@ -89,14 +89,14 @@ public class TestAddIndexes extends Luce
assertEquals(40, writer.maxDoc());
writer.close();
- writer = newWriter(aux2, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE));
+ writer = newWriter(aux2, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE));
// add 50 documents in compound files
addDocs2(writer, 50);
assertEquals(50, writer.maxDoc());
writer.close();
// test doc count before segments are merged
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
assertEquals(100, writer.maxDoc());
writer.addIndexes(aux, aux2);
assertEquals(190, writer.maxDoc());
@@ -111,14 +111,14 @@ public class TestAddIndexes extends Luce
// now add another set in.
Directory aux3 = newDirectory();
- writer = newWriter(aux3, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = newWriter(aux3, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
// add 40 documents
addDocs(writer, 40);
assertEquals(40, writer.maxDoc());
writer.close();
// test doc count before segments are merged
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
assertEquals(190, writer.maxDoc());
writer.addIndexes(aux3);
assertEquals(230, writer.maxDoc());
@@ -132,7 +132,7 @@ public class TestAddIndexes extends Luce
verifyTermDocs(dir, new Term("content", "bbb"), 50);
// now fully merge it.
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
writer.forceMerge(1);
writer.close();
@@ -145,11 +145,11 @@ public class TestAddIndexes extends Luce
// now add a single document
Directory aux4 = newDirectory();
- writer = newWriter(aux4, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = newWriter(aux4, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
addDocs2(writer, 1);
writer.close();
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
assertEquals(230, writer.maxDoc());
writer.addIndexes(aux4);
assertEquals(231, writer.maxDoc());
@@ -172,7 +172,7 @@ public class TestAddIndexes extends Luce
Directory aux = newDirectory();
setUpDirs(dir, aux);
- IndexWriter writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ IndexWriter writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
writer.addIndexes(aux);
// Adds 10 docs, then replaces them with another 10
@@ -208,7 +208,7 @@ public class TestAddIndexes extends Luce
Directory aux = newDirectory();
setUpDirs(dir, aux);
- IndexWriter writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ IndexWriter writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
// Adds 10 docs, then replaces them with another 10
// docs, so 10 pending deletes:
@@ -246,7 +246,7 @@ public class TestAddIndexes extends Luce
Directory aux = newDirectory();
setUpDirs(dir, aux);
- IndexWriter writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ IndexWriter writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
// Adds 10 docs, then replaces them with another 10
// docs, so 10 pending deletes:
@@ -286,7 +286,7 @@ public class TestAddIndexes extends Luce
IndexWriter writer = null;
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
// add 100 documents
addDocs(writer, 100);
assertEquals(100, writer.maxDoc());
@@ -294,7 +294,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
aux,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setMaxBufferedDocs(1000).
setMergePolicy(newLogMergePolicy(false))
@@ -304,7 +304,7 @@ public class TestAddIndexes extends Luce
writer.close();
writer = newWriter(
aux,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setMaxBufferedDocs(1000).
setMergePolicy(newLogMergePolicy(false))
@@ -312,7 +312,7 @@ public class TestAddIndexes extends Luce
addDocs(writer, 100);
writer.close();
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
try {
// cannot add self
writer.addIndexes(aux, dir);
@@ -342,7 +342,7 @@ public class TestAddIndexes extends Luce
IndexWriter writer = newWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setMaxBufferedDocs(10).
setMergePolicy(newLogMergePolicy(4))
@@ -371,7 +371,7 @@ public class TestAddIndexes extends Luce
IndexWriter writer = newWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setMaxBufferedDocs(9).
setMergePolicy(newLogMergePolicy(4))
@@ -400,13 +400,13 @@ public class TestAddIndexes extends Luce
IndexWriter writer = newWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setMaxBufferedDocs(10).
setMergePolicy(newLogMergePolicy(4))
);
- writer.addIndexes(aux, new MockDirectoryWrapper(random, new RAMDirectory(aux, newIOContext(random))));
+ writer.addIndexes(aux, new MockDirectoryWrapper(random(), new RAMDirectory(aux, newIOContext(random()))));
assertEquals(1060, writer.maxDoc());
assertEquals(1000, writer.getDocCount(0));
writer.close();
@@ -426,7 +426,7 @@ public class TestAddIndexes extends Luce
setUpDirs(dir, aux, true);
- IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
IndexWriter writer = new IndexWriter(aux, dontMergeConfig);
for (int i = 0; i < 20; i++) {
@@ -439,7 +439,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setMaxBufferedDocs(4).
setMergePolicy(newLogMergePolicy(4))
@@ -448,7 +448,7 @@ public class TestAddIndexes extends Luce
if (VERBOSE) {
System.out.println("\nTEST: now addIndexes");
}
- writer.addIndexes(aux, new MockDirectoryWrapper(random, new RAMDirectory(aux, newIOContext(random))));
+ writer.addIndexes(aux, new MockDirectoryWrapper(random(), new RAMDirectory(aux, newIOContext(random()))));
assertEquals(1020, writer.maxDoc());
assertEquals(1000, writer.getDocCount(0));
writer.close();
@@ -468,7 +468,7 @@ public class TestAddIndexes extends Luce
IndexWriter writer = newWriter(
aux2,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setMaxBufferedDocs(100).
setMergePolicy(newLogMergePolicy(10))
@@ -478,7 +478,7 @@ public class TestAddIndexes extends Luce
assertEquals(3, writer.getSegmentCount());
writer.close();
- IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
writer = new IndexWriter(aux, dontMergeConfig);
for (int i = 0; i < 27; i++) {
@@ -489,7 +489,7 @@ public class TestAddIndexes extends Luce
assertEquals(3, reader.numDocs());
reader.close();
- dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
writer = new IndexWriter(aux2, dontMergeConfig);
for (int i = 0; i < 8; i++) {
@@ -502,7 +502,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setMaxBufferedDocs(6).
setMergePolicy(newLogMergePolicy(4))
@@ -550,7 +550,7 @@ public class TestAddIndexes extends Luce
private void verifyTermDocs(Directory dir, Term term, int numDocs)
throws IOException {
IndexReader reader = IndexReader.open(dir);
- DocsEnum docsEnum = _TestUtil.docs(random, reader, term.field, term.bytes, null, null, false);
+ DocsEnum docsEnum = _TestUtil.docs(random(), reader, term.field, term.bytes, null, null, false);
int count = 0;
while (docsEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS)
count++;
@@ -565,7 +565,7 @@ public class TestAddIndexes extends Luce
private void setUpDirs(Directory dir, Directory aux, boolean withID) throws IOException {
IndexWriter writer = null;
- writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE).setMaxBufferedDocs(1000));
+ writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE).setMaxBufferedDocs(1000));
// add 1000 documents in 1 segment
if (withID) {
addDocsWithID(writer, 1000, 0);
@@ -578,7 +578,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
aux,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setMaxBufferedDocs(1000).
setMergePolicy(newLogMergePolicy(false, 10))
@@ -593,7 +593,7 @@ public class TestAddIndexes extends Luce
writer.close();
writer = newWriter(
aux,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setMaxBufferedDocs(1000).
setMergePolicy(newLogMergePolicy(false, 10))
@@ -612,7 +612,7 @@ public class TestAddIndexes extends Luce
lmp.setUseCompoundFile(false);
lmp.setMergeFactor(100);
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(5).setMergePolicy(lmp));
Document doc = new Document();
@@ -641,7 +641,7 @@ public class TestAddIndexes extends Luce
lmp.setUseCompoundFile(false);
lmp.setMergeFactor(4);
writer = new IndexWriter(dir2, newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random))
+ new MockAnalyzer(random()))
.setMergeScheduler(new SerialMergeScheduler()).setMergePolicy(lmp));
writer.addIndexes(dir);
writer.close();
@@ -672,16 +672,16 @@ public class TestAddIndexes extends Luce
public RunAddIndexesThreads(int numCopy) throws Throwable {
NUM_COPY = numCopy;
- dir = new MockDirectoryWrapper(random, new RAMDirectory());
+ dir = new MockDirectoryWrapper(random(), new RAMDirectory());
IndexWriter writer = new IndexWriter(dir, new IndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(2));
for (int i = 0; i < NUM_INIT_DOCS; i++)
addDoc(writer);
writer.close();
dir2 = newDirectory();
- writer2 = new IndexWriter(dir2, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer2 = new IndexWriter(dir2, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
writer2.commit();
@@ -700,7 +700,7 @@ public class TestAddIndexes extends Luce
final Directory[] dirs = new Directory[NUM_COPY];
for(int k=0;k<NUM_COPY;k++)
- dirs[k] = new MockDirectoryWrapper(random, new RAMDirectory(dir, newIOContext(random)));
+ dirs[k] = new MockDirectoryWrapper(random(), new RAMDirectory(dir, newIOContext(random())));
int j=0;
@@ -925,7 +925,7 @@ public class TestAddIndexes extends Luce
CommitAndAddIndexes3 c = new CommitAndAddIndexes3(NUM_COPY);
c.launchThreads(-1);
- Thread.sleep(_TestUtil.nextInt(random, 10, 500));
+ Thread.sleep(_TestUtil.nextInt(random(), 10, 500));
// Close w/o first stopping/joining the threads
if (VERBOSE) {
@@ -950,7 +950,7 @@ public class TestAddIndexes extends Luce
CommitAndAddIndexes3 c = new CommitAndAddIndexes3(NUM_COPY);
c.launchThreads(-1);
- Thread.sleep(_TestUtil.nextInt(random, 10, 500));
+ Thread.sleep(_TestUtil.nextInt(random(), 10, 500));
// Close w/o first stopping/joining the threads
if (VERBOSE) {
@@ -971,7 +971,7 @@ public class TestAddIndexes extends Luce
Directory[] dirs = new Directory[2];
for (int i = 0; i < dirs.length; i++) {
dirs[i] = newDirectory();
- IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
IndexWriter writer = new IndexWriter(dirs[i], conf);
Document doc = new Document();
doc.add(new StringField("id", "myid"));
@@ -979,7 +979,7 @@ public class TestAddIndexes extends Luce
writer.close();
}
- IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
IndexWriter writer = new IndexWriter(dirs[0], conf);
// Now delete the document
@@ -1020,7 +1020,7 @@ public class TestAddIndexes extends Luce
IndexWriter writer = null;
writer = newWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE).setCodec(codec));
+ new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE).setCodec(codec));
// add 100 documents
addDocsWithID(writer, 100, 0);
assertEquals(100, writer.maxDoc());
@@ -1030,7 +1030,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
aux,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setCodec(codec).
setMaxBufferedDocs(10).
@@ -1044,7 +1044,7 @@ public class TestAddIndexes extends Luce
writer = newWriter(
aux2,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setCodec(codec)
);
@@ -1057,7 +1057,7 @@ public class TestAddIndexes extends Luce
// test doc count before segments are merged
writer = newWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.APPEND).
setCodec(codec)
);
@@ -1094,7 +1094,7 @@ public class TestAddIndexes extends Luce
Directory[] dirs = new Directory[2];
for (int i = 0; i < dirs.length; i++) {
dirs[i] = new RAMDirectory();
- IndexWriter w = new IndexWriter(dirs[i], new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter w = new IndexWriter(dirs[i], new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document d = new Document();
FieldType customType = new FieldType(TextField.TYPE_STORED);
customType.setStoreTermVectors(true);
@@ -1105,8 +1105,8 @@ public class TestAddIndexes extends Luce
IndexReader[] readers = new IndexReader[] { IndexReader.open(dirs[0]), IndexReader.open(dirs[1]) };
- Directory dir = new MockDirectoryWrapper(random, new RAMDirectory());
- IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy());
+ Directory dir = new MockDirectoryWrapper(random(), new RAMDirectory());
+ IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(newLogMergePolicy());
LogMergePolicy lmp = (LogMergePolicy) conf.getMergePolicy();
lmp.setUseCompoundFile(true);
lmp.setNoCFSRatio(1.0); // Force creation of CFS
@@ -1175,7 +1175,7 @@ public class TestAddIndexes extends Luce
toAdd.setCheckIndexOnClose(false);
{
IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random));
+ new MockAnalyzer(random()));
conf.setCodec(new UnRegisteredCodec());
IndexWriter w = new IndexWriter(toAdd, conf);
Document doc = new Document();
@@ -1189,8 +1189,8 @@ public class TestAddIndexes extends Luce
{
Directory dir = newDirectory();
IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random));
- conf.setCodec(_TestUtil.alwaysPostingsFormat(new Pulsing40PostingsFormat(1 + random.nextInt(20))));
+ new MockAnalyzer(random()));
+ conf.setCodec(_TestUtil.alwaysPostingsFormat(new Pulsing40PostingsFormat(1 + random().nextInt(20))));
IndexWriter w = new IndexWriter(dir, conf);
try {
w.addIndexes(toAdd);
@@ -1217,7 +1217,7 @@ public class TestAddIndexes extends Luce
// LUCENE-3575
public void testFieldNamesChanged() throws IOException {
Directory d1 = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, d1);
+ RandomIndexWriter w = new RandomIndexWriter(random(), d1);
Document doc = new Document();
doc.add(newField("f1", "doc1 field1", StringField.TYPE_STORED));
doc.add(newField("id", "1", StringField.TYPE_STORED));
@@ -1226,7 +1226,7 @@ public class TestAddIndexes extends Luce
w.close();
Directory d2 = newDirectory();
- w = new RandomIndexWriter(random, d2);
+ w = new RandomIndexWriter(random(), d2);
doc = new Document();
doc.add(newField("f2", "doc2 field2", StringField.TYPE_STORED));
doc.add(newField("id", "2", StringField.TYPE_STORED));
@@ -1235,7 +1235,7 @@ public class TestAddIndexes extends Luce
w.close();
Directory d3 = newDirectory();
- w = new RandomIndexWriter(random, d3);
+ w = new RandomIndexWriter(random(), d3);
w.addIndexes(r1, r2);
r1.close();
d1.close();
@@ -1260,7 +1260,7 @@ public class TestAddIndexes extends Luce
public void testDocValues() throws IOException {
assumeFalse("preflex does not support docvalues", Codec.getDefault().getName().equals("Lucene3x"));
Directory d1 = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, d1);
+ RandomIndexWriter w = new RandomIndexWriter(random(), d1);
Document doc = new Document();
doc.add(newField("id", "1", StringField.TYPE_STORED));
doc.add(new DocValuesField("dv", 1, DocValues.Type.VAR_INTS));
@@ -1269,7 +1269,7 @@ public class TestAddIndexes extends Luce
w.close();
Directory d2 = newDirectory();
- w = new RandomIndexWriter(random, d2);
+ w = new RandomIndexWriter(random(), d2);
doc = new Document();
doc.add(newField("id", "2", StringField.TYPE_STORED));
doc.add(new DocValuesField("dv", 2, DocValues.Type.VAR_INTS));
@@ -1278,7 +1278,7 @@ public class TestAddIndexes extends Luce
w.close();
Directory d3 = newDirectory();
- w = new RandomIndexWriter(random, d3);
+ w = new RandomIndexWriter(random(), d3);
w.addIndexes(SlowCompositeReaderWrapper.wrap(r1), SlowCompositeReaderWrapper.wrap(r2));
r1.close();
d1.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAtomicUpdate.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAtomicUpdate.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAtomicUpdate.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestAtomicUpdate.java Sun Apr 15 14:41:44 2012
@@ -16,29 +16,23 @@ package org.apache.lucene.index;
* limitations under the License.
*/
-import org.apache.lucene.util.*;
-import org.apache.lucene.store.*;
-import org.apache.lucene.document.*;
-import org.apache.lucene.analysis.MockAnalyzer;
-
-import java.util.Random;
import java.io.File;
import java.io.IOException;
+import org.apache.lucene.analysis.MockAnalyzer;
+import org.apache.lucene.document.*;
+import org.apache.lucene.store.*;
+import org.apache.lucene.util.*;
+
public class TestAtomicUpdate extends LuceneTestCase {
-
private static final class MockIndexWriter extends IndexWriter {
-
- static Random RANDOM;
-
public MockIndexWriter(Directory dir, IndexWriterConfig conf) throws IOException {
super(dir, conf);
}
@Override
boolean testPoint(String name) {
- // if (name.equals("startCommit")) {
- if (RANDOM.nextInt(4) == 2)
+ if (LuceneTestCase.random().nextInt(4) == 2)
Thread.yield();
return true;
}
@@ -127,7 +121,7 @@ public class TestAtomicUpdate extends Lu
TimedThread[] threads = new TimedThread[4];
IndexWriterConfig conf = new IndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(7);
((TieredMergePolicy) conf.getMergePolicy()).setMaxMergeAtOnce(3);
IndexWriter writer = new MockIndexWriter(directory, conf);
@@ -185,11 +179,10 @@ public class TestAtomicUpdate extends Lu
FSDirectory.
*/
public void testAtomicUpdates() throws Exception {
- MockIndexWriter.RANDOM = random;
Directory directory;
// First in a RAM directory:
- directory = new MockDirectoryWrapper(random, new RAMDirectory());
+ directory = new MockDirectoryWrapper(random(), new RAMDirectory());
runTest(directory);
directory.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java Sun Apr 15 14:41:44 2012
@@ -176,7 +176,7 @@ public class TestBackwardsCompatibility
try {
writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
fail("IndexWriter creation should not pass for "+unsupportedNames[i]);
} catch (IndexFormatTooOldException e) {
// pass
@@ -217,7 +217,7 @@ public class TestBackwardsCompatibility
}
Directory dir = newDirectory(oldIndexDirs.get(name));
IndexWriter w = new IndexWriter(dir, new IndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
w.forceMerge(1);
w.close();
@@ -229,7 +229,7 @@ public class TestBackwardsCompatibility
for (String name : oldNames) {
Directory targetDir = newDirectory();
IndexWriter w = new IndexWriter(targetDir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
w.addIndexes(oldIndexDirs.get(name));
w.close();
@@ -243,7 +243,7 @@ public class TestBackwardsCompatibility
Directory targetDir = newDirectory();
IndexWriter w = new IndexWriter(targetDir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
w.addIndexes(reader);
w.close();
reader.close();
@@ -261,7 +261,7 @@ public class TestBackwardsCompatibility
public void testIndexOldIndexNoAdds() throws IOException {
for (String name : oldNames) {
Directory dir = newDirectory(oldIndexDirs.get(name));
- changeIndexNoAdds(random, dir);
+ changeIndexNoAdds(random(), dir);
dir.close();
}
}
@@ -272,7 +272,7 @@ public class TestBackwardsCompatibility
System.out.println("TEST: oldName=" + name);
}
Directory dir = newDirectory(oldIndexDirs.get(name));
- changeIndexWithAdds(random, dir, name);
+ changeIndexWithAdds(random(), dir, name);
dir.close();
}
}
@@ -426,7 +426,7 @@ public class TestBackwardsCompatibility
mp.setUseCompoundFile(doCFS);
mp.setNoCFSRatio(1.0);
// TODO: remove randomness
- IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(10).setMergePolicy(mp);
IndexWriter writer = new IndexWriter(dir, conf);
@@ -445,7 +445,7 @@ public class TestBackwardsCompatibility
mp.setUseCompoundFile(doCFS);
mp.setNoCFSRatio(1.0);
// TODO: remove randomness
- conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(10).setMergePolicy(mp);
writer = new IndexWriter(dir, conf);
addNoProxDoc(writer);
@@ -480,7 +480,7 @@ public class TestBackwardsCompatibility
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(-1).
setRAMBufferSizeMB(16.0).
setMergePolicy(mergePolicy)
@@ -494,7 +494,7 @@ public class TestBackwardsCompatibility
// Delete one doc so we get a .del file:
writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.NO_COMPOUND_FILES)
);
Term searchTerm = new Term("id", "7");
@@ -597,7 +597,7 @@ public class TestBackwardsCompatibility
// should be found exactly
assertEquals(TermsEnum.SeekStatus.FOUND,
terms.seekCeil(aaaTerm));
- assertEquals(35, countDocs(_TestUtil.docs(random, terms, null, null, false)));
+ assertEquals(35, countDocs(_TestUtil.docs(random(), terms, null, null, false)));
assertNull(terms.next());
// should hit end of field
@@ -609,12 +609,12 @@ public class TestBackwardsCompatibility
assertEquals(TermsEnum.SeekStatus.NOT_FOUND,
terms.seekCeil(new BytesRef("a")));
assertTrue(terms.term().bytesEquals(aaaTerm));
- assertEquals(35, countDocs(_TestUtil.docs(random, terms, null, null, false)));
+ assertEquals(35, countDocs(_TestUtil.docs(random(), terms, null, null, false)));
assertNull(terms.next());
assertEquals(TermsEnum.SeekStatus.FOUND,
terms.seekCeil(aaaTerm));
- assertEquals(35, countDocs(_TestUtil.docs(random, terms,null, null, false)));
+ assertEquals(35, countDocs(_TestUtil.docs(random(), terms,null, null, false)));
assertNull(terms.next());
r.close();
@@ -713,12 +713,12 @@ public class TestBackwardsCompatibility
RAMDirectory ramDir = new RAMDirectory();
for (int i = 0; i < 3; i++) {
// only use Log- or TieredMergePolicy, to make document addition predictable and not suddenly merge:
- MergePolicy mp = random.nextBoolean() ? newLogMergePolicy() : newTieredMergePolicy();
- IndexWriterConfig iwc = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ MergePolicy mp = random().nextBoolean() ? newLogMergePolicy() : newTieredMergePolicy();
+ IndexWriterConfig iwc = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(mp);
IndexWriter w = new IndexWriter(ramDir, iwc);
// add few more docs:
- for(int j = 0; j < RANDOM_MULTIPLIER * random.nextInt(30); j++) {
+ for(int j = 0; j < RANDOM_MULTIPLIER * random().nextInt(30); j++) {
addDoc(w, id++);
}
w.close(false);
@@ -726,7 +726,7 @@ public class TestBackwardsCompatibility
// add dummy segments (which are all in current
// version) to single segment index
- MergePolicy mp = random.nextBoolean() ? newLogMergePolicy() : newTieredMergePolicy();
+ MergePolicy mp = random().nextBoolean() ? newLogMergePolicy() : newTieredMergePolicy();
IndexWriterConfig iwc = new IndexWriterConfig(TEST_VERSION_CURRENT, null)
.setMergePolicy(mp);
IndexWriter w = new IndexWriter(dir, iwc);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBinaryTerms.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBinaryTerms.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBinaryTerms.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestBinaryTerms.java Sun Apr 15 14:41:44 2012
@@ -40,7 +40,7 @@ public class TestBinaryTerms extends Luc
Codec.getDefault().getName().equals("Lucene3x"));
Directory dir = newDirectory();
- RandomIndexWriter iw = new RandomIndexWriter(random, dir);
+ RandomIndexWriter iw = new RandomIndexWriter(random(), dir);
BytesRef bytes = new BytesRef(2);
BinaryTokenStream tokenStream = new BinaryTokenStream(bytes);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestByteSlices.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestByteSlices.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestByteSlices.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestByteSlices.java Sun Apr 15 14:41:44 2012
@@ -42,7 +42,7 @@ public class TestByteSlices extends Luce
int num = atLeast(10000);
for (int iter = 0; iter < num; iter++) {
- int stream = random.nextInt(NUM_STREAM);
+ int stream = random().nextInt(NUM_STREAM);
if (VERBOSE)
System.out.println("write stream=" + stream);
@@ -54,12 +54,12 @@ public class TestByteSlices extends Luce
}
writer.init(uptos[stream]);
- int numValue = random.nextInt(20);
+ int numValue = random().nextInt(20);
for(int j=0;j<numValue;j++) {
if (VERBOSE)
System.out.println(" write " + (counters[stream]+j));
// write some large (incl. negative) ints:
- writer.writeVInt(random.nextInt());
+ writer.writeVInt(random().nextInt());
writer.writeVInt(counters[stream]+j);
}
counters[stream] += numValue;
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCheckIndex.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCheckIndex.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCheckIndex.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCheckIndex.java Sun Apr 15 14:41:44 2012
@@ -35,7 +35,7 @@ public class TestCheckIndex extends Luce
public void testDeletedDocs() throws IOException {
Directory dir = newDirectory();
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2));
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMaxBufferedDocs(2));
for(int i=0;i<19;i++) {
Document doc = new Document();
FieldType customType = new FieldType(TextField.TYPE_STORED);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCodecs.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCodecs.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCodecs.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestCodecs.java Sun Apr 15 14:41:44 2012
@@ -179,7 +179,7 @@ public class TestCodecs extends LuceneTe
final private static String SEGMENT = "0";
TermData[] makeRandomTerms(final boolean omitTF, final boolean storePayloads) {
- final int numTerms = 1+random.nextInt(NUM_TERMS_RAND);
+ final int numTerms = 1+random().nextInt(NUM_TERMS_RAND);
//final int numTerms = 2;
final TermData[] terms = new TermData[numTerms];
@@ -190,14 +190,14 @@ public class TestCodecs extends LuceneTe
// Make term text
String text2;
while(true) {
- text2 = _TestUtil.randomUnicodeString(random);
+ text2 = _TestUtil.randomUnicodeString(random());
if (!termsSeen.contains(text2) && !text2.endsWith(".")) {
termsSeen.add(text2);
break;
}
}
- final int docFreq = 1+random.nextInt(DOC_FREQ_RAND);
+ final int docFreq = 1+random().nextInt(DOC_FREQ_RAND);
final int[] docs = new int[docFreq];
PositionData[][] positions;
@@ -208,21 +208,21 @@ public class TestCodecs extends LuceneTe
int docID = 0;
for(int j=0;j<docFreq;j++) {
- docID += _TestUtil.nextInt(random, 1, 10);
+ docID += _TestUtil.nextInt(random(), 1, 10);
docs[j] = docID;
if (!omitTF) {
- final int termFreq = 1+random.nextInt(TERM_DOC_FREQ_RAND);
+ final int termFreq = 1+random().nextInt(TERM_DOC_FREQ_RAND);
positions[j] = new PositionData[termFreq];
int position = 0;
for(int k=0;k<termFreq;k++) {
- position += _TestUtil.nextInt(random, 1, 10);
+ position += _TestUtil.nextInt(random(), 1, 10);
final BytesRef payload;
- if (storePayloads && random.nextInt(4) == 0) {
- final byte[] bytes = new byte[1+random.nextInt(5)];
+ if (storePayloads && random().nextInt(4) == 0) {
+ final byte[] bytes = new byte[1+random().nextInt(5)];
for(int l=0;l<bytes.length;l++) {
- bytes[l] = (byte) random.nextInt(255);
+ bytes[l] = (byte) random().nextInt(255);
}
payload = new BytesRef(bytes);
} else {
@@ -260,7 +260,7 @@ public class TestCodecs extends LuceneTe
Codec codec = Codec.getDefault();
final SegmentInfo si = new SegmentInfo(SEGMENT, 10000, dir, false, codec, clonedFieldInfos);
- final FieldsProducer reader = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR));
+ final FieldsProducer reader = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random()), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR));
final FieldsEnum fieldsEnum = reader.iterator();
assertNotNull(fieldsEnum.next());
@@ -279,7 +279,7 @@ public class TestCodecs extends LuceneTe
// make sure it properly fully resets (rewinds) its
// internal state:
for(int iter=0;iter<2;iter++) {
- docsEnum = _TestUtil.docs(random, termsEnum, null, docsEnum, false);
+ docsEnum = _TestUtil.docs(random(), termsEnum, null, docsEnum, false);
assertEquals(terms[i].docs[0], docsEnum.nextDoc());
assertEquals(DocIdSetIterator.NO_MORE_DOCS, docsEnum.nextDoc());
}
@@ -319,7 +319,7 @@ public class TestCodecs extends LuceneTe
if (VERBOSE) {
System.out.println("TEST: now read postings");
}
- final FieldsProducer terms = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR));
+ final FieldsProducer terms = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random()), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR));
final Verify[] threads = new Verify[NUM_TEST_THREADS-1];
for(int i=0;i<NUM_TEST_THREADS-1;i++) {
@@ -342,7 +342,7 @@ public class TestCodecs extends LuceneTe
public void testSepPositionAfterMerge() throws IOException {
final Directory dir = newDirectory();
final IndexWriterConfig config = newIndexWriterConfig(Version.LUCENE_31,
- new MockAnalyzer(random));
+ new MockAnalyzer(random()));
config.setCodec(_TestUtil.alwaysPostingsFormat(new MockSepPostingsFormat()));
final IndexWriter writer = new IndexWriter(dir, config);
@@ -439,7 +439,7 @@ public class TestCodecs extends LuceneTe
assertEquals(positions[i].pos, pos);
if (positions[i].payload != null) {
assertTrue(posEnum.hasPayload());
- if (LuceneTestCase.random.nextInt(3) < 2) {
+ if (random().nextInt(3) < 2) {
// Verify the payload bytes
final BytesRef otherPayload = posEnum.getPayload();
assertTrue("expected=" + positions[i].payload.toString() + " got=" + otherPayload.toString(), positions[i].payload.equals(otherPayload));
@@ -453,7 +453,7 @@ public class TestCodecs extends LuceneTe
public void _run() throws Throwable {
for(int iter=0;iter<NUM_TEST_ITER;iter++) {
- final FieldData field = fields[LuceneTestCase.random.nextInt(fields.length)];
+ final FieldData field = fields[random().nextInt(fields.length)];
final TermsEnum termsEnum = termsDict.terms(field.fieldInfo.name).iterator(null);
if (si.getCodec() instanceof Lucene3xCodec) {
// code below expects unicode sort order
@@ -473,18 +473,18 @@ public class TestCodecs extends LuceneTe
assertEquals(upto, field.terms.length);
// Test random seek:
- TermData term = field.terms[LuceneTestCase.random.nextInt(field.terms.length)];
+ TermData term = field.terms[random().nextInt(field.terms.length)];
TermsEnum.SeekStatus status = termsEnum.seekCeil(new BytesRef(term.text2));
assertEquals(status, TermsEnum.SeekStatus.FOUND);
assertEquals(term.docs.length, termsEnum.docFreq());
if (field.omitTF) {
- this.verifyDocs(term.docs, term.positions, _TestUtil.docs(random, termsEnum, null, null, false), false);
+ this.verifyDocs(term.docs, term.positions, _TestUtil.docs(random(), termsEnum, null, null, false), false);
} else {
this.verifyDocs(term.docs, term.positions, termsEnum.docsAndPositions(null, null, false), true);
}
// Test random seek by ord:
- final int idx = LuceneTestCase.random.nextInt(field.terms.length);
+ final int idx = random().nextInt(field.terms.length);
term = field.terms[idx];
boolean success = false;
try {
@@ -498,7 +498,7 @@ public class TestCodecs extends LuceneTe
assertTrue(termsEnum.term().bytesEquals(new BytesRef(term.text2)));
assertEquals(term.docs.length, termsEnum.docFreq());
if (field.omitTF) {
- this.verifyDocs(term.docs, term.positions, _TestUtil.docs(random, termsEnum, null, null, false), false);
+ this.verifyDocs(term.docs, term.positions, _TestUtil.docs(random(), termsEnum, null, null, false), false);
} else {
this.verifyDocs(term.docs, term.positions, termsEnum.docsAndPositions(null, null, false), true);
}
@@ -509,7 +509,7 @@ public class TestCodecs extends LuceneTe
System.out.println("TEST: seek non-exist terms");
}
for(int i=0;i<100;i++) {
- final String text2 = _TestUtil.randomUnicodeString(random) + ".";
+ final String text2 = _TestUtil.randomUnicodeString(random()) + ".";
status = termsEnum.seekCeil(new BytesRef(text2));
assertTrue(status == TermsEnum.SeekStatus.NOT_FOUND ||
status == TermsEnum.SeekStatus.END);
@@ -547,7 +547,7 @@ public class TestCodecs extends LuceneTe
upto = 0;
do {
term = field.terms[upto];
- if (LuceneTestCase.random.nextInt(3) == 1) {
+ if (random().nextInt(3) == 1) {
final DocsEnum docs;
final DocsEnum docsAndFreqs;
final DocsAndPositionsEnum postings;
@@ -556,12 +556,12 @@ public class TestCodecs extends LuceneTe
if (postings != null) {
docs = docsAndFreqs = postings;
} else {
- docs = docsAndFreqs = _TestUtil.docs(random, termsEnum, null, null, true);
+ docs = docsAndFreqs = _TestUtil.docs(random(), termsEnum, null, null, true);
}
} else {
postings = null;
docsAndFreqs = null;
- docs = _TestUtil.docs(random, termsEnum, null, null, false);
+ docs = _TestUtil.docs(random(), termsEnum, null, null, false);
}
assertNotNull(docs);
int upto2 = -1;
@@ -569,10 +569,10 @@ public class TestCodecs extends LuceneTe
// Maybe skip:
final int left = term.docs.length-upto2;
int doc;
- if (LuceneTestCase.random.nextInt(3) == 1 && left >= 1) {
- final int inc = 1+LuceneTestCase.random.nextInt(left-1);
+ if (random().nextInt(3) == 1 && left >= 1) {
+ final int inc = 1+random().nextInt(left-1);
upto2 += inc;
- if (LuceneTestCase.random.nextInt(2) == 1) {
+ if (random().nextInt(2) == 1) {
doc = docs.advance(term.docs[upto2]);
assertEquals(term.docs[upto2], doc);
} else {
@@ -597,7 +597,7 @@ public class TestCodecs extends LuceneTe
assertEquals(term.docs[upto2], doc);
if (!field.omitTF) {
assertEquals(term.positions[upto2].length, postings.freq());
- if (LuceneTestCase.random.nextInt(2) == 1) {
+ if (random().nextInt(2) == 1) {
this.verifyPositions(term.positions[upto2], postings);
}
}
@@ -616,9 +616,9 @@ public class TestCodecs extends LuceneTe
private void write(final FieldInfos fieldInfos, final Directory dir, final FieldData[] fields, boolean allowPreFlex) throws Throwable {
- final int termIndexInterval = _TestUtil.nextInt(random, 13, 27);
+ final int termIndexInterval = _TestUtil.nextInt(random(), 13, 27);
final Codec codec = Codec.getDefault();
- final SegmentWriteState state = new SegmentWriteState(InfoStream.getDefault(), dir, SEGMENT, fieldInfos, 10000, termIndexInterval, codec, null, newIOContext(random));
+ final SegmentWriteState state = new SegmentWriteState(InfoStream.getDefault(), dir, SEGMENT, fieldInfos, 10000, termIndexInterval, codec, null, newIOContext(random()));
final FieldsConsumer consumer = codec.postingsFormat().fieldsConsumer(state);
Arrays.sort(fields);