You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by dw...@apache.org on 2012/04/15 16:42:01 UTC
svn commit: r1326351 [8/22] - in /lucene/dev/trunk: ./ dev-tools/eclipse/
lucene/
lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/
lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/custom/
lucene/contrib/high...
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelAtomicReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelAtomicReader.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelAtomicReader.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelAtomicReader.java Sun Apr 15 14:41:44 2012
@@ -34,8 +34,8 @@ public class TestParallelAtomicReader ex
private Directory dir, dir1, dir2;
public void testQueries() throws Exception {
- single = single(random);
- parallel = parallel(random);
+ single = single(random());
+ parallel = parallel(random());
queryTest(new TermQuery(new Term("f1", "v1")));
queryTest(new TermQuery(new Term("f1", "v2")));
@@ -59,8 +59,8 @@ public class TestParallelAtomicReader ex
}
public void testFieldNames() throws Exception {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
ParallelAtomicReader pr = new ParallelAtomicReader(SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir1)),
SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir2)));
FieldInfos fieldInfos = pr.getFieldInfos();
@@ -75,8 +75,8 @@ public class TestParallelAtomicReader ex
}
public void testRefCounts1() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
AtomicReader ir1, ir2;
// close subreaders, ParallelReader will not change refCounts, but close on its own close
ParallelAtomicReader pr = new ParallelAtomicReader(ir1 = SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir1)),
@@ -93,8 +93,8 @@ public class TestParallelAtomicReader ex
}
public void testRefCounts2() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
AtomicReader ir1 = SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir1));
AtomicReader ir2 = SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir2));
// don't close subreaders, so ParallelReader will increment refcounts
@@ -115,11 +115,11 @@ public class TestParallelAtomicReader ex
public void testIncompatibleIndexes() throws IOException {
// two documents:
- Directory dir1 = getDir1(random);
+ Directory dir1 = getDir1(random());
// one document only:
Directory dir2 = newDirectory();
- IndexWriter w2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter w2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document d3 = new Document();
d3.add(newField("f3", "v1", TextField.TYPE_STORED));
@@ -137,7 +137,7 @@ public class TestParallelAtomicReader ex
}
try {
- new ParallelAtomicReader(random.nextBoolean(),
+ new ParallelAtomicReader(random().nextBoolean(),
new AtomicReader[] {ir1, ir2},
new AtomicReader[] {ir1, ir2});
fail("didn't get expected exception: indexes don't have same number of documents");
@@ -154,8 +154,8 @@ public class TestParallelAtomicReader ex
}
public void testIgnoreStoredFields() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
AtomicReader ir1 = SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir1));
AtomicReader ir2 = SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir2));
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelCompositeReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelCompositeReader.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelCompositeReader.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelCompositeReader.java Sun Apr 15 14:41:44 2012
@@ -34,8 +34,8 @@ public class TestParallelCompositeReader
private Directory dir, dir1, dir2;
public void testQueries() throws Exception {
- single = single(random, false);
- parallel = parallel(random, false);
+ single = single(random(), false);
+ parallel = parallel(random(), false);
queries();
@@ -47,8 +47,8 @@ public class TestParallelCompositeReader
}
public void testQueriesCompositeComposite() throws Exception {
- single = single(random, true);
- parallel = parallel(random, true);
+ single = single(random(), true);
+ parallel = parallel(random(), true);
queries();
@@ -76,8 +76,8 @@ public class TestParallelCompositeReader
}
public void testRefCounts1() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
DirectoryReader ir1, ir2;
// close subreaders, ParallelReader will not change refCounts, but close on its own close
ParallelCompositeReader pr = new ParallelCompositeReader(ir1 = DirectoryReader.open(dir1),
@@ -93,8 +93,8 @@ public class TestParallelCompositeReader
}
public void testRefCounts2() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
DirectoryReader ir1 = DirectoryReader.open(dir1);
DirectoryReader ir2 = DirectoryReader.open(dir2);
@@ -116,11 +116,11 @@ public class TestParallelCompositeReader
public void testIncompatibleIndexes1() throws IOException {
// two documents:
- Directory dir1 = getDir1(random);
+ Directory dir1 = getDir1(random());
// one document only:
Directory dir2 = newDirectory();
- IndexWriter w2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter w2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document d3 = new Document();
d3.add(newField("f3", "v1", TextField.TYPE_STORED));
@@ -136,7 +136,7 @@ public class TestParallelCompositeReader
// expected exception
}
try {
- new ParallelCompositeReader(random.nextBoolean(), ir1, ir2);
+ new ParallelCompositeReader(random().nextBoolean(), ir1, ir2);
fail("didn't get expected exception: indexes don't have same number of documents");
} catch (IllegalArgumentException e) {
// expected exception
@@ -152,8 +152,8 @@ public class TestParallelCompositeReader
}
public void testIncompatibleIndexes2() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getInvalidStructuredDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getInvalidStructuredDir2(random());
DirectoryReader ir1 = DirectoryReader.open(dir1),
ir2 = DirectoryReader.open(dir2);
@@ -165,7 +165,7 @@ public class TestParallelCompositeReader
// expected exception
}
try {
- new ParallelCompositeReader(random.nextBoolean(), readers, readers);
+ new ParallelCompositeReader(random().nextBoolean(), readers, readers);
fail("didn't get expected exception: indexes don't have same subreader structure");
} catch (IllegalArgumentException e) {
// expected exception
@@ -181,8 +181,8 @@ public class TestParallelCompositeReader
}
public void testIncompatibleIndexes3() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
CompositeReader ir1 = new MultiReader(DirectoryReader.open(dir1), SlowCompositeReaderWrapper.wrap(DirectoryReader.open(dir1))),
ir2 = new MultiReader(DirectoryReader.open(dir2), DirectoryReader.open(dir2));
@@ -194,7 +194,7 @@ public class TestParallelCompositeReader
// expected exception
}
try {
- new ParallelCompositeReader(random.nextBoolean(), readers, readers);
+ new ParallelCompositeReader(random().nextBoolean(), readers, readers);
fail("didn't get expected exception: indexes don't have same subreader structure");
} catch (IllegalArgumentException e) {
// expected exception
@@ -210,8 +210,8 @@ public class TestParallelCompositeReader
}
public void testIgnoreStoredFields() throws IOException {
- Directory dir1 = getDir1(random);
- Directory dir2 = getDir2(random);
+ Directory dir1 = getDir1(random());
+ Directory dir2 = getDir2(random());
CompositeReader ir1 = DirectoryReader.open(dir1);
CompositeReader ir2 = DirectoryReader.open(dir2);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java Sun Apr 15 14:41:44 2012
@@ -42,14 +42,14 @@ public class TestParallelReaderEmptyInde
*/
public void testEmptyIndex() throws IOException {
Directory rd1 = newDirectory();
- IndexWriter iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
iw.close();
// create a copy:
Directory rd2 = newDirectory(rd1);
Directory rdOut = newDirectory();
- IndexWriter iwOut = new IndexWriter(rdOut, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iwOut = new IndexWriter(rdOut, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
ParallelAtomicReader apr = new ParallelAtomicReader(
SlowCompositeReaderWrapper.wrap(DirectoryReader.open(rd1)),
@@ -89,7 +89,7 @@ public class TestParallelReaderEmptyInde
public void testEmptyIndexWithVectors() throws IOException {
Directory rd1 = newDirectory();
{
- IndexWriter iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document doc = new Document();
Field idField = newField("id", "", TextField.TYPE_UNSTORED);
doc.add(idField);
@@ -103,7 +103,7 @@ public class TestParallelReaderEmptyInde
iw.addDocument(doc);
iw.close();
- IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
IndexWriter writer = new IndexWriter(rd1, dontMergeConfig);
@@ -114,14 +114,14 @@ public class TestParallelReaderEmptyInde
assertEquals(1, ir.numDocs());
ir.close();
- iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
iw.forceMerge(1);
iw.close();
}
Directory rd2 = newDirectory();
{
- IndexWriter iw = new IndexWriter(rd2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iw = new IndexWriter(rd2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document doc = new Document();
iw.addDocument(doc);
iw.close();
@@ -129,7 +129,7 @@ public class TestParallelReaderEmptyInde
Directory rdOut = newDirectory();
- IndexWriter iwOut = new IndexWriter(rdOut, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iwOut = new IndexWriter(rdOut, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
final DirectoryReader reader1, reader2;
ParallelAtomicReader pr = new ParallelAtomicReader(
SlowCompositeReaderWrapper.wrap(reader1 = DirectoryReader.open(rd1)),
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java Sun Apr 15 14:41:44 2012
@@ -40,7 +40,8 @@ public class TestParallelTermEnum extend
super.setUp();
Document doc;
rd1 = newDirectory();
- IndexWriter iw1 = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iw1 = new IndexWriter(rd1, newIndexWriterConfig(
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
doc = new Document();
doc.add(newField("field1", "the quick brown fox jumps", TextField.TYPE_STORED));
@@ -49,7 +50,8 @@ public class TestParallelTermEnum extend
iw1.close();
rd2 = newDirectory();
- IndexWriter iw2 = new IndexWriter(rd2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter iw2 = new IndexWriter(rd2, newIndexWriterConfig(
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
doc = new Document();
doc.add(newField("field1", "the fox jumps over the lazy dog", TextField.TYPE_STORED));
@@ -79,7 +81,7 @@ public class TestParallelTermEnum extend
BytesRef b = te.next();
assertNotNull(b);
assertEquals(t, b.utf8ToString());
- DocsEnum td = _TestUtil.docs(random, te, liveDocs, null, false);
+ DocsEnum td = _TestUtil.docs(random(), te, liveDocs, null, false);
assertTrue(td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS);
assertEquals(0, td.docID());
assertEquals(td.nextDoc(), DocIdSetIterator.NO_MORE_DOCS);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java Sun Apr 15 14:41:44 2012
@@ -221,25 +221,25 @@ public class TestPayloadProcessorProvide
@Test
public void testAddIndexes() throws Exception {
// addIndexes - single commit in each
- doTest(random, true, 0, false);
+ doTest(random(), true, 0, false);
// addIndexes - multiple commits in each
- doTest(random, true, 0, true);
+ doTest(random(), true, 0, true);
}
@Test
public void testAddIndexesIntoExisting() throws Exception {
// addIndexes - single commit in each
- doTest(random, false, NUM_DOCS, false);
+ doTest(random(), false, NUM_DOCS, false);
// addIndexes - multiple commits in each
- doTest(random, false, NUM_DOCS, true);
+ doTest(random(), false, NUM_DOCS, true);
}
@Test
public void testRegularMerges() throws Exception {
Directory dir = newDirectory();
- populateDocs(random, dir, true);
+ populateDocs(random(), dir, true);
verifyPayloadExists(dir, "p", new BytesRef("p1"), NUM_DOCS);
verifyPayloadExists(dir, "p", new BytesRef("p2"), NUM_DOCS);
@@ -247,7 +247,7 @@ public class TestPayloadProcessorProvide
// won't get processed.
Map<Directory, ReaderPayloadProcessor> processors = new HashMap<Directory, ReaderPayloadProcessor>();
processors.put(dir, new PerTermPayloadProcessor());
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false)));
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random(), MockTokenizer.WHITESPACE, false)));
writer.setPayloadProcessorProvider(new PerDirPayloadProcessor(processors));
writer.forceMerge(1);
writer.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloads.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloads.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloads.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPayloads.java Sun Apr 15 14:41:44 2012
@@ -243,7 +243,7 @@ public class TestPayloads extends Lucene
// handle a caller that mucks with the
// returned payload:
if (rarely()) {
- br.bytes = new byte[random.nextInt(5)];
+ br.bytes = new byte[random().nextInt(5)];
}
br.length = 0;
br.offset = 0;
@@ -351,7 +351,7 @@ public class TestPayloads extends Lucene
static final Charset utf8 = Charset.forName("UTF-8");
private void generateRandomData(byte[] data) {
// this test needs the random data to be valid unicode
- String s = _TestUtil.randomFixedByteLengthUnicodeString(random, data.length);
+ String s = _TestUtil.randomFixedByteLengthUnicodeString(random(), data.length);
byte b[] = s.getBytes(utf8);
assert b.length == data.length;
System.arraycopy(b, 0, data, 0, b.length);
@@ -503,7 +503,7 @@ public class TestPayloads extends Lucene
Directory dir = newDirectory();
final IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
final String field = "test";
Thread[] ingesters = new Thread[numThreads];
@@ -612,15 +612,15 @@ public class TestPayloads extends Lucene
public void testAcrossFields() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir,
- new MockAnalyzer(random, MockTokenizer.WHITESPACE, true));
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir,
+ new MockAnalyzer(random(), MockTokenizer.WHITESPACE, true));
Document doc = new Document();
doc.add(new Field("hasMaybepayload", "here we go", TextField.TYPE_STORED));
writer.addDocument(doc);
writer.close();
- writer = new RandomIndexWriter(random, dir,
- new MockAnalyzer(random, MockTokenizer.WHITESPACE, true));
+ writer = new RandomIndexWriter(random(), dir,
+ new MockAnalyzer(random(), MockTokenizer.WHITESPACE, true));
doc = new Document();
doc.add(new Field("hasMaybepayload2", "here we go", TextField.TYPE_STORED));
writer.addDocument(doc);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java Sun Apr 15 14:41:44 2012
@@ -37,9 +37,9 @@ import org.apache.lucene.util._TestUtil;
public class TestPerSegmentDeletes extends LuceneTestCase {
public void testDeletes1() throws Exception {
//IndexWriter.debug2 = System.out;
- Directory dir = new MockDirectoryWrapper(new Random(random.nextLong()), new RAMDirectory());
+ Directory dir = new MockDirectoryWrapper(new Random(random().nextLong()), new RAMDirectory());
IndexWriterConfig iwc = new IndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random));
+ new MockAnalyzer(random()));
iwc.setMergeScheduler(new SerialMergeScheduler());
iwc.setMaxBufferedDocs(5000);
iwc.setRAMBufferSizeMB(100);
@@ -220,13 +220,13 @@ public class TestPerSegmentDeletes exten
}
}
- public static int[] toDocsArray(Term term, Bits bits, IndexReader reader)
+ public int[] toDocsArray(Term term, Bits bits, IndexReader reader)
throws IOException {
Fields fields = MultiFields.getFields(reader);
Terms cterms = fields.terms(term.field);
TermsEnum ctermsEnum = cterms.iterator(null);
if (ctermsEnum.seekExact(new BytesRef(term.text()), false)) {
- DocsEnum docsEnum = _TestUtil.docs(random, ctermsEnum, bits, null, false);
+ DocsEnum docsEnum = _TestUtil.docs(random(), ctermsEnum, bits, null, false);
return toArray(docsEnum);
}
return null;
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPersistentSnapshotDeletionPolicy.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPersistentSnapshotDeletionPolicy.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPersistentSnapshotDeletionPolicy.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPersistentSnapshotDeletionPolicy.java Sun Apr 15 14:41:44 2012
@@ -79,7 +79,7 @@ public class TestPersistentSnapshotDelet
int numSnapshots = 3;
Directory dir = newDirectory();
PersistentSnapshotDeletionPolicy psdp = (PersistentSnapshotDeletionPolicy) getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, psdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), psdp));
prepareIndexAndSnapshots(psdp, writer, numSnapshots, "snapshot");
writer.close();
psdp.close();
@@ -88,7 +88,7 @@ public class TestPersistentSnapshotDelet
psdp = new PersistentSnapshotDeletionPolicy(
new KeepOnlyLastCommitDeletionPolicy(), snapshotDir, OpenMode.APPEND,
TEST_VERSION_CURRENT);
- new IndexWriter(dir, getConfig(random, psdp)).close();
+ new IndexWriter(dir, getConfig(random(), psdp)).close();
assertSnapshotExists(dir, psdp, numSnapshots);
assertEquals(numSnapshots, psdp.getSnapshots().size());
@@ -104,7 +104,7 @@ public class TestPersistentSnapshotDelet
@Test
public void testInvalidSnapshotInfos() throws Exception {
// Add the correct number of documents (1), but without snapshot information
- IndexWriter writer = new IndexWriter(snapshotDir, getConfig(random, null));
+ IndexWriter writer = new IndexWriter(snapshotDir, getConfig(random(), null));
writer.addDocument(new Document());
writer.close();
try {
@@ -119,7 +119,7 @@ public class TestPersistentSnapshotDelet
@Test
public void testNoSnapshotInfos() throws Exception {
// Initialize an empty index in snapshotDir - PSDP should initialize successfully.
- new IndexWriter(snapshotDir, getConfig(random, null)).close();
+ new IndexWriter(snapshotDir, getConfig(random(), null)).close();
new PersistentSnapshotDeletionPolicy(
new KeepOnlyLastCommitDeletionPolicy(), snapshotDir, OpenMode.APPEND,
TEST_VERSION_CURRENT).close();
@@ -128,7 +128,7 @@ public class TestPersistentSnapshotDelet
@Test(expected=IllegalStateException.class)
public void testTooManySnapshotInfos() throws Exception {
// Write two documents to the snapshots directory - illegal.
- IndexWriter writer = new IndexWriter(snapshotDir, getConfig(random, null));
+ IndexWriter writer = new IndexWriter(snapshotDir, getConfig(random(), null));
writer.addDocument(new Document());
writer.addDocument(new Document());
writer.close();
@@ -143,7 +143,7 @@ public class TestPersistentSnapshotDelet
public void testSnapshotRelease() throws Exception {
Directory dir = newDirectory();
PersistentSnapshotDeletionPolicy psdp = (PersistentSnapshotDeletionPolicy) getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, psdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), psdp));
prepareIndexAndSnapshots(psdp, writer, 1, "snapshot");
writer.close();
@@ -166,7 +166,7 @@ public class TestPersistentSnapshotDelet
int numSnapshots = 1;
Directory dir = newDirectory();
PersistentSnapshotDeletionPolicy psdp = (PersistentSnapshotDeletionPolicy) getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, psdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), psdp));
prepareIndexAndSnapshots(psdp, writer, numSnapshots, "snapshot");
writer.close();
dir.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java Sun Apr 15 14:41:44 2012
@@ -53,11 +53,11 @@ public class TestPostingsOffsets extends
super.setUp();
// Currently only SimpleText and Lucene40 can index offsets into postings:
assumeTrue("codec does not support offsets", Codec.getDefault().getName().equals("SimpleText") || Codec.getDefault().getName().equals("Lucene40"));
- iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
if (Codec.getDefault().getName().equals("Lucene40")) {
// pulsing etc are not implemented
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
iwc.setCodec(_TestUtil.alwaysPostingsFormat(new Lucene40PostingsFormat()));
} else {
iwc.setCodec(_TestUtil.alwaysPostingsFormat(new MemoryPostingsFormat()));
@@ -68,7 +68,7 @@ public class TestPostingsOffsets extends
public void testBasic() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, dir, iwc);
+ RandomIndexWriter w = new RandomIndexWriter(random(), dir, iwc);
Document doc = new Document();
FieldType ft = new FieldType(TextField.TYPE_UNSTORED);
@@ -129,25 +129,25 @@ public class TestPostingsOffsets extends
public void doTestNumbers(boolean withPayloads) throws Exception {
Directory dir = newDirectory();
- Analyzer analyzer = withPayloads ? new MockPayloadAnalyzer() : new MockAnalyzer(random);
+ Analyzer analyzer = withPayloads ? new MockPayloadAnalyzer() : new MockAnalyzer(random());
iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer);
if (Codec.getDefault().getName().equals("Lucene40")) {
// pulsing etc are not implemented
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
iwc.setCodec(_TestUtil.alwaysPostingsFormat(new Lucene40PostingsFormat()));
} else {
iwc.setCodec(_TestUtil.alwaysPostingsFormat(new MemoryPostingsFormat()));
}
}
iwc.setMergePolicy(newLogMergePolicy()); // will rely on docids a bit for skipping
- RandomIndexWriter w = new RandomIndexWriter(random, dir, iwc);
+ RandomIndexWriter w = new RandomIndexWriter(random(), dir, iwc);
FieldType ft = new FieldType(TextField.TYPE_STORED);
ft.setIndexOptions(FieldInfo.IndexOptions.DOCS_AND_FREQS_AND_POSITIONS_AND_OFFSETS);
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
ft.setStoreTermVectors(true);
- ft.setStoreTermVectorOffsets(random.nextBoolean());
- ft.setStoreTermVectorPositions(random.nextBoolean());
+ ft.setStoreTermVectorOffsets(random().nextBoolean());
+ ft.setStoreTermVectorPositions(random().nextBoolean());
}
int numDocs = atLeast(500);
@@ -192,7 +192,7 @@ public class TestPostingsOffsets extends
int numSkippingTests = atLeast(50);
for (int j = 0; j < numSkippingTests; j++) {
- int num = _TestUtil.nextInt(random, 100, Math.min(numDocs-1, 999));
+ int num = _TestUtil.nextInt(random(), 100, Math.min(numDocs-1, 999));
DocsAndPositionsEnum dp = MultiFields.getTermPositionsEnum(reader, null, "numbers", new BytesRef("hundred"), true);
int doc = dp.advance(num);
assertEquals(num, doc);
@@ -232,7 +232,7 @@ public class TestPostingsOffsets extends
final Map<String,Map<Integer,List<Token>>> actualTokens = new HashMap<String,Map<Integer,List<Token>>>();
Directory dir = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, dir, iwc);
+ RandomIndexWriter w = new RandomIndexWriter(random(), dir, iwc);
final int numDocs = atLeast(20);
//final int numDocs = atLeast(5);
@@ -242,10 +242,10 @@ public class TestPostingsOffsets extends
// TODO: randomize what IndexOptions we use; also test
// changing this up in one IW buffered segment...:
ft.setIndexOptions(FieldInfo.IndexOptions.DOCS_AND_FREQS_AND_POSITIONS_AND_OFFSETS);
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
ft.setStoreTermVectors(true);
- ft.setStoreTermVectorOffsets(random.nextBoolean());
- ft.setStoreTermVectorPositions(random.nextBoolean());
+ ft.setStoreTermVectorOffsets(random().nextBoolean());
+ ft.setStoreTermVectorPositions(random().nextBoolean());
}
for(int docCount=0;docCount<numDocs;docCount++) {
@@ -259,22 +259,22 @@ public class TestPostingsOffsets extends
//System.out.println("doc id=" + docCount);
for(int tokenCount=0;tokenCount<numTokens;tokenCount++) {
final String text;
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
text = "a";
- } else if (random.nextBoolean()) {
+ } else if (random().nextBoolean()) {
text = "b";
- } else if (random.nextBoolean()) {
+ } else if (random().nextBoolean()) {
text = "c";
} else {
text = "d";
}
- int posIncr = random.nextBoolean() ? 1 : random.nextInt(5);
+ int posIncr = random().nextBoolean() ? 1 : random().nextInt(5);
if (tokenCount == 0 && posIncr == 0) {
posIncr = 1;
}
- final int offIncr = random.nextBoolean() ? 0 : random.nextInt(5);
- final int tokenOffset = random.nextInt(5);
+ final int offIncr = random().nextBoolean() ? 0 : random().nextInt(5);
+ final int tokenOffset = random().nextInt(5);
final Token token = makeToken(text, posIncr, offset+offIncr, offset+offIncr+tokenOffset);
if (!actualTokens.containsKey(text)) {
@@ -310,7 +310,7 @@ public class TestPostingsOffsets extends
final int docIDToID[] = FieldCache.DEFAULT.getInts(sub, "id", false);
for(String term : terms) {
//System.out.println(" term=" + term);
- if (termsEnum.seekExact(new BytesRef(term), random.nextBoolean())) {
+ if (termsEnum.seekExact(new BytesRef(term), random().nextBoolean())) {
docs = termsEnum.docs(null, docs, true);
assertNotNull(docs);
int doc;
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPrefixCodedTerms.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPrefixCodedTerms.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPrefixCodedTerms.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestPrefixCodedTerms.java Sun Apr 15 14:41:44 2012
@@ -49,7 +49,7 @@ public class TestPrefixCodedTerms extend
Set<Term> terms = new TreeSet<Term>();
int nterms = atLeast(10000);
for (int i = 0; i < nterms; i++) {
- Term term = new Term(_TestUtil.randomUnicodeString(random, 2), _TestUtil.randomUnicodeString(random));
+ Term term = new Term(_TestUtil.randomUnicodeString(random(), 2), _TestUtil.randomUnicodeString(random()));
terms.add(term);
}
@@ -101,14 +101,14 @@ public class TestPrefixCodedTerms extend
}
public void testMergeRandom() {
- PrefixCodedTerms pb[] = new PrefixCodedTerms[_TestUtil.nextInt(random, 2, 10)];
+ PrefixCodedTerms pb[] = new PrefixCodedTerms[_TestUtil.nextInt(random(), 2, 10)];
Set<Term> superSet = new TreeSet<Term>();
for (int i = 0; i < pb.length; i++) {
Set<Term> terms = new TreeSet<Term>();
- int nterms = _TestUtil.nextInt(random, 0, 10000);
+ int nterms = _TestUtil.nextInt(random(), 0, 10000);
for (int j = 0; j < nterms; j++) {
- Term term = new Term(_TestUtil.randomUnicodeString(random, 2), _TestUtil.randomUnicodeString(random, 4));
+ Term term = new Term(_TestUtil.randomUnicodeString(random(), 2), _TestUtil.randomUnicodeString(random(), 4));
terms.add(term);
}
superSet.addAll(terms);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRandomStoredFields.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRandomStoredFields.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRandomStoredFields.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRandomStoredFields.java Sun Apr 15 14:41:44 2012
@@ -41,8 +41,8 @@ public class TestRandomStoredFields exte
public void testRandomStoredFields() throws IOException {
Directory dir = newDirectory();
- Random rand = random;
- RandomIndexWriter w = new RandomIndexWriter(rand, dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(_TestUtil.nextInt(rand, 5, 20)));
+ Random rand = random();
+ RandomIndexWriter w = new RandomIndexWriter(rand, dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMaxBufferedDocs(_TestUtil.nextInt(rand, 5, 20)));
//w.w.setUseCompoundFile(false);
final int docCount = atLeast(200);
final int fieldCount = _TestUtil.nextInt(rand, 1, 5);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestReaderClosed.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestReaderClosed.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestReaderClosed.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestReaderClosed.java Sun Apr 15 14:41:44 2012
@@ -37,9 +37,9 @@ public class TestReaderClosed extends Lu
public void setUp() throws Exception {
super.setUp();
dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.KEYWORD, false))
- .setMaxBufferedDocs(_TestUtil.nextInt(random, 50, 1000)));
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random(), MockTokenizer.KEYWORD, false))
+ .setMaxBufferedDocs(_TestUtil.nextInt(random(), 50, 1000)));
Document doc = new Document();
Field field = newField("field", "", StringField.TYPE_UNSTORED);
@@ -49,7 +49,7 @@ public class TestReaderClosed extends Lu
// but for preflex codec, the test can be very slow, so use less iterations.
int num = atLeast(10);
for (int i = 0; i < num; i++) {
- field.setStringValue(_TestUtil.randomUnicodeString(random, 10));
+ field.setStringValue(_TestUtil.randomUnicodeString(random(), 10));
writer.addDocument(doc);
}
reader = writer.getReader();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollback.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollback.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollback.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollback.java Sun Apr 15 14:41:44 2012
@@ -28,7 +28,7 @@ public class TestRollback extends Lucene
// LUCENE-2536
public void testRollbackIntegrityWithBufferFlush() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter rw = new RandomIndexWriter(random, dir);
+ RandomIndexWriter rw = new RandomIndexWriter(random(), dir);
for (int i = 0; i < 5; i++) {
Document doc = new Document();
doc.add(newField("pk", Integer.toString(i), StringField.TYPE_STORED));
@@ -37,7 +37,7 @@ public class TestRollback extends Lucene
rw.close();
// If buffer size is small enough to cause a flush, errors ensue...
- IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2).setOpenMode(IndexWriterConfig.OpenMode.APPEND));
+ IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMaxBufferedDocs(2).setOpenMode(IndexWriterConfig.OpenMode.APPEND));
for (int i = 0; i < 3; i++) {
Document doc = new Document();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollingUpdates.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollingUpdates.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollingUpdates.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestRollingUpdates.java Sun Apr 15 14:41:44 2012
@@ -17,6 +17,8 @@ package org.apache.lucene.index;
* limitations under the License.
*/
+import java.util.Random;
+
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.codecs.Codec;
import org.apache.lucene.codecs.memory.MemoryPostingsFormat;
@@ -32,20 +34,21 @@ public class TestRollingUpdates extends
@Test
public void testRollingUpdates() throws Exception {
+ Random random = new Random(random().nextLong());
final MockDirectoryWrapper dir = newDirectory();
dir.setCheckIndexOnClose(false); // we use a custom codec provider
final LineFileDocs docs = new LineFileDocs(random, defaultCodecSupportsDocValues());
//provider.register(new MemoryCodec());
- if ( (!"Lucene3x".equals(Codec.getDefault().getName())) && random.nextBoolean()) {
- Codec.setDefault(_TestUtil.alwaysPostingsFormat(new MemoryPostingsFormat(random.nextBoolean())));
+ if ( (!"Lucene3x".equals(Codec.getDefault().getName())) && random().nextBoolean()) {
+ Codec.setDefault(_TestUtil.alwaysPostingsFormat(new MemoryPostingsFormat(random().nextBoolean())));
}
- final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
final int SIZE = atLeast(20);
int id = 0;
IndexReader r = null;
- final int numUpdates = (int) (SIZE * (2+(TEST_NIGHTLY ? 200*random.nextDouble() : 5*random.nextDouble())));
+ final int numUpdates = (int) (SIZE * (2+(TEST_NIGHTLY ? 200*random().nextDouble() : 5*random().nextDouble())));
if (VERBOSE) {
System.out.println("TEST: numUpdates=" + numUpdates);
}
@@ -60,11 +63,11 @@ public class TestRollingUpdates extends
((Field) doc.getField("docid")).setStringValue(myID);
w.updateDocument(new Term("docid", myID), doc);
- if (docIter >= SIZE && random.nextInt(50) == 17) {
+ if (docIter >= SIZE && random().nextInt(50) == 17) {
if (r != null) {
r.close();
}
- final boolean applyDeletions = random.nextBoolean();
+ final boolean applyDeletions = random().nextBoolean();
r = w.getReader(applyDeletions);
assertTrue("applyDeletions=" + applyDeletions + " r.numDocs()=" + r.numDocs() + " vs SIZE=" + SIZE, !applyDeletions || r.numDocs() == SIZE);
}
@@ -88,12 +91,12 @@ public class TestRollingUpdates extends
public void testUpdateSameDoc() throws Exception {
final Directory dir = newDirectory();
- final LineFileDocs docs = new LineFileDocs(random);
+ final LineFileDocs docs = new LineFileDocs(random());
for (int r = 0; r < 3; r++) {
final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMaxBufferedDocs(2));
final int numUpdates = atLeast(20);
- int numThreads = _TestUtil.nextInt(random, 2, 6);
+ int numThreads = _TestUtil.nextInt(random(), 2, 6);
IndexingThread[] threads = new IndexingThread[numThreads];
for (int i = 0; i < numThreads; i++) {
threads[i] = new IndexingThread(docs, w, numUpdates);
@@ -133,7 +136,7 @@ public class TestRollingUpdates extends
Document doc = new Document();// docs.nextDoc();
doc.add(newField("id", "test", StringField.TYPE_UNSTORED));
writer.updateDocument(new Term("id", "test"), doc);
- if (random.nextInt(3) == 0) {
+ if (random().nextInt(3) == 0) {
if (open == null) {
open = IndexReader.open(writer, true);
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSameTokenSamePosition.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSameTokenSamePosition.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSameTokenSamePosition.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSameTokenSamePosition.java Sun Apr 15 14:41:44 2012
@@ -36,7 +36,7 @@ public class TestSameTokenSamePosition e
*/
public void test() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter riw = new RandomIndexWriter(random, dir);
+ RandomIndexWriter riw = new RandomIndexWriter(random(), dir);
Document doc = new Document();
doc.add(new TextField("eng", new BugReproTokenStream()));
riw.addDocument(doc);
@@ -49,7 +49,7 @@ public class TestSameTokenSamePosition e
*/
public void testMoreDocs() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter riw = new RandomIndexWriter(random, dir);
+ RandomIndexWriter riw = new RandomIndexWriter(random(), dir);
for (int i = 0; i < 100; i++) {
Document doc = new Document();
doc.add(new TextField("eng", new BugReproTokenStream()));
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java Sun Apr 15 14:41:44 2012
@@ -49,11 +49,11 @@ public class TestSegmentMerger extends L
merge1Dir = newDirectory();
merge2Dir = newDirectory();
DocHelper.setupDoc(doc1);
- SegmentInfo info1 = DocHelper.writeDoc(random, merge1Dir, doc1);
+ SegmentInfo info1 = DocHelper.writeDoc(random(), merge1Dir, doc1);
DocHelper.setupDoc(doc2);
- SegmentInfo info2 = DocHelper.writeDoc(random, merge2Dir, doc2);
- reader1 = new SegmentReader(info1, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random));
- reader2 = new SegmentReader(info2, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random));
+ SegmentInfo info2 = DocHelper.writeDoc(random(), merge2Dir, doc2);
+ reader1 = new SegmentReader(info1, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random()));
+ reader2 = new SegmentReader(info2, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random()));
}
@Override
@@ -76,7 +76,7 @@ public class TestSegmentMerger extends L
public void testMerge() throws IOException {
final Codec codec = Codec.getDefault();
- SegmentMerger merger = new SegmentMerger(InfoStream.getDefault(), mergedDir, IndexWriterConfig.DEFAULT_TERM_INDEX_INTERVAL, mergedSegment, MergeState.CheckAbort.NONE, null, new FieldInfos(new FieldInfos.FieldNumberBiMap()), codec, newIOContext(random));
+ SegmentMerger merger = new SegmentMerger(InfoStream.getDefault(), mergedDir, IndexWriterConfig.DEFAULT_TERM_INDEX_INTERVAL, mergedSegment, MergeState.CheckAbort.NONE, null, new FieldInfos(new FieldInfos.FieldNumberBiMap()), codec, newIOContext(random()));
merger.add(reader1);
merger.add(reader2);
MergeState mergeState = merger.merge();
@@ -86,7 +86,7 @@ public class TestSegmentMerger extends L
//Should be able to open a new SegmentReader against the new directory
SegmentReader mergedReader = new SegmentReader(new SegmentInfo(mergedSegment, docsMerged, mergedDir, false,
codec, fieldInfos),
- DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random));
+ DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random()));
assertTrue(mergedReader != null);
assertTrue(mergedReader.numDocs() == 2);
Document newDoc1 = mergedReader.document(0);
@@ -97,7 +97,7 @@ public class TestSegmentMerger extends L
assertTrue(newDoc2 != null);
assertTrue(DocHelper.numFields(newDoc2) == DocHelper.numFields(doc2) - DocHelper.unstored.size());
- DocsEnum termDocs = _TestUtil.docs(random, mergedReader,
+ DocsEnum termDocs = _TestUtil.docs(random(), mergedReader,
DocHelper.TEXT_FIELD_2_KEY,
new BytesRef("field"),
MultiFields.getLiveDocs(mergedReader),
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java Sun Apr 15 14:41:44 2012
@@ -41,7 +41,7 @@ public class TestSegmentReader extends L
super.setUp();
dir = newDirectory();
DocHelper.setupDoc(testDoc);
- SegmentInfo info = DocHelper.writeDoc(random, dir, testDoc);
+ SegmentInfo info = DocHelper.writeDoc(random(), dir, testDoc);
reader = new SegmentReader(info, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, IOContext.READ);
}
@@ -128,7 +128,7 @@ public class TestSegmentReader extends L
}
}
- DocsEnum termDocs = _TestUtil.docs(random, reader,
+ DocsEnum termDocs = _TestUtil.docs(random(), reader,
DocHelper.TEXT_FIELD_1_KEY,
new BytesRef("field"),
MultiFields.getLiveDocs(reader),
@@ -136,7 +136,7 @@ public class TestSegmentReader extends L
false);
assertTrue(termDocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS);
- termDocs = _TestUtil.docs(random, reader,
+ termDocs = _TestUtil.docs(random(), reader,
DocHelper.NO_NORMS_KEY,
new BytesRef(DocHelper.NO_NORMS_TEXT),
MultiFields.getLiveDocs(reader),
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java Sun Apr 15 14:41:44 2012
@@ -38,7 +38,7 @@ public class TestSegmentTermDocs extends
super.setUp();
dir = newDirectory();
DocHelper.setupDoc(testDoc);
- info = DocHelper.writeDoc(random, dir, testDoc);
+ info = DocHelper.writeDoc(random(), dir, testDoc);
}
@Override
@@ -57,13 +57,13 @@ public class TestSegmentTermDocs extends
public void testTermDocs(int indexDivisor) throws IOException {
//After adding the document, we should be able to read it back in
- SegmentReader reader = new SegmentReader(info, indexDivisor, newIOContext(random));
+ SegmentReader reader = new SegmentReader(info, indexDivisor, newIOContext(random()));
assertTrue(reader != null);
assertEquals(indexDivisor, reader.getTermInfosIndexDivisor());
TermsEnum terms = reader.fields().terms(DocHelper.TEXT_FIELD_2_KEY).iterator(null);
terms.seekCeil(new BytesRef("field"));
- DocsEnum termDocs = _TestUtil.docs(random, terms, reader.getLiveDocs(), null, true);
+ DocsEnum termDocs = _TestUtil.docs(random(), terms, reader.getLiveDocs(), null, true);
if (termDocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) {
int docId = termDocs.docID();
assertTrue(docId == 0);
@@ -80,9 +80,9 @@ public class TestSegmentTermDocs extends
public void testBadSeek(int indexDivisor) throws IOException {
{
//After adding the document, we should be able to read it back in
- SegmentReader reader = new SegmentReader(info, indexDivisor, newIOContext(random));
+ SegmentReader reader = new SegmentReader(info, indexDivisor, newIOContext(random()));
assertTrue(reader != null);
- DocsEnum termDocs = _TestUtil.docs(random, reader,
+ DocsEnum termDocs = _TestUtil.docs(random(), reader,
"textField2",
new BytesRef("bad"),
reader.getLiveDocs(),
@@ -94,9 +94,9 @@ public class TestSegmentTermDocs extends
}
{
//After adding the document, we should be able to read it back in
- SegmentReader reader = new SegmentReader(info, indexDivisor, newIOContext(random));
+ SegmentReader reader = new SegmentReader(info, indexDivisor, newIOContext(random()));
assertTrue(reader != null);
- DocsEnum termDocs = _TestUtil.docs(random, reader,
+ DocsEnum termDocs = _TestUtil.docs(random(), reader,
"junk",
new BytesRef("bad"),
reader.getLiveDocs(),
@@ -113,7 +113,7 @@ public class TestSegmentTermDocs extends
public void testSkipTo(int indexDivisor) throws IOException {
Directory dir = newDirectory();
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy()));
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(newLogMergePolicy()));
Term ta = new Term("content","aaa");
for(int i = 0; i < 10; i++)
@@ -133,7 +133,7 @@ public class TestSegmentTermDocs extends
IndexReader reader = IndexReader.open(dir, indexDivisor);
- DocsEnum tdocs = _TestUtil.docs(random, reader,
+ DocsEnum tdocs = _TestUtil.docs(random(), reader,
ta.field(),
new BytesRef(ta.text()),
MultiFields.getLiveDocs(reader),
@@ -158,7 +158,7 @@ public class TestSegmentTermDocs extends
assertFalse(tdocs.advance(10) != DocIdSetIterator.NO_MORE_DOCS);
// without next
- tdocs = _TestUtil.docs(random, reader,
+ tdocs = _TestUtil.docs(random(), reader,
ta.field(),
new BytesRef(ta.text()),
MultiFields.getLiveDocs(reader),
@@ -176,7 +176,7 @@ public class TestSegmentTermDocs extends
// exactly skipInterval documents and therefore with optimization
// with next
- tdocs = _TestUtil.docs(random, reader,
+ tdocs = _TestUtil.docs(random(), reader,
tb.field(),
new BytesRef(tb.text()),
MultiFields.getLiveDocs(reader),
@@ -200,7 +200,7 @@ public class TestSegmentTermDocs extends
assertFalse(tdocs.advance(26) != DocIdSetIterator.NO_MORE_DOCS);
// without next
- tdocs = _TestUtil.docs(random, reader,
+ tdocs = _TestUtil.docs(random(), reader,
tb.field(),
new BytesRef(tb.text()),
MultiFields.getLiveDocs(reader),
@@ -220,7 +220,7 @@ public class TestSegmentTermDocs extends
// much more than skipInterval documents and therefore with optimization
// with next
- tdocs = _TestUtil.docs(random, reader,
+ tdocs = _TestUtil.docs(random(), reader,
tc.field(),
new BytesRef(tc.text()),
MultiFields.getLiveDocs(reader),
@@ -246,7 +246,7 @@ public class TestSegmentTermDocs extends
assertFalse(tdocs.advance(76) != DocIdSetIterator.NO_MORE_DOCS);
//without next
- tdocs = _TestUtil.docs(random, reader,
+ tdocs = _TestUtil.docs(random(), reader,
tc.field(),
new BytesRef(tc.text()),
MultiFields.getLiveDocs(reader),
@@ -271,7 +271,7 @@ public class TestSegmentTermDocs extends
public void testIndexDivisor() throws IOException {
testDoc = new Document();
DocHelper.setupDoc(testDoc);
- DocHelper.writeDoc(random, dir, testDoc);
+ DocHelper.writeDoc(random(), dir, testDoc);
testTermDocs(2);
testBadSeek(2);
testSkipTo(2);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermEnum.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermEnum.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermEnum.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSegmentTermEnum.java Sun Apr 15 14:41:44 2012
@@ -49,7 +49,7 @@ public class TestSegmentTermEnum extends
public void testTermEnum() throws IOException {
IndexWriter writer = null;
- writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
// ADD 100 documents with term : aaa
// add 100 documents with terms: aaa bbb
@@ -65,7 +65,7 @@ public class TestSegmentTermEnum extends
verifyDocFreq();
// merge segments
- writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND));
+ writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND));
writer.forceMerge(1);
writer.close();
@@ -75,7 +75,7 @@ public class TestSegmentTermEnum extends
public void testPrevTermAtEnd() throws IOException
{
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setCodec(_TestUtil.alwaysPostingsFormat(new Lucene40PostingsFormat())));
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setCodec(_TestUtil.alwaysPostingsFormat(new Lucene40PostingsFormat())));
addDoc(writer, "aaa bbb");
writer.close();
SegmentReader reader = getOnlySegmentReader(IndexReader.open(dir));
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java Sun Apr 15 14:41:44 2012
@@ -92,7 +92,7 @@ public class TestSnapshotDeletionPolicy
@Test
public void testSnapshotDeletionPolicy() throws Exception {
Directory fsDir = newDirectory();
- runTest(random, fsDir);
+ runTest(random(), fsDir);
fsDir.close();
}
@@ -206,7 +206,7 @@ public class TestSnapshotDeletionPolicy
byte[] buffer = new byte[4096];
private void readFile(Directory dir, String name) throws Exception {
- IndexInput input = dir.openInput(name, newIOContext(random));
+ IndexInput input = dir.openInput(name, newIOContext(random()));
try {
long size = dir.fileLength(name);
long bytesLeft = size;
@@ -238,7 +238,7 @@ public class TestSnapshotDeletionPolicy
// Create 3 snapshots: snapshot0, snapshot1, snapshot2
Directory dir = newDirectory();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
prepareIndexAndSnapshots(sdp, writer, numSnapshots, "snapshot");
writer.close();
@@ -249,7 +249,7 @@ public class TestSnapshotDeletionPolicy
// open a new IndexWriter w/ no snapshots to keep and assert that all snapshots are gone.
sdp = getDeletionPolicy();
- writer = new IndexWriter(dir, getConfig(random, sdp));
+ writer = new IndexWriter(dir, getConfig(random(), sdp));
writer.deleteUnusedFiles();
writer.close();
assertEquals("no snapshots should exist", 1, DirectoryReader.listCommits(dir).size());
@@ -269,7 +269,7 @@ public class TestSnapshotDeletionPolicy
public void testMultiThreadedSnapshotting() throws Exception {
Directory dir = newDirectory();
final SnapshotDeletionPolicy sdp = getDeletionPolicy();
- final IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ final IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
Thread[] threads = new Thread[10];
for (int i = 0; i < threads.length; i++) {
@@ -314,12 +314,12 @@ public class TestSnapshotDeletionPolicy
int numSnapshots = 2;
Directory dir = newDirectory();
SnapshotDeletionPolicy sdp = getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
prepareIndexAndSnapshots(sdp, writer, numSnapshots, "snapshot");
writer.close();
// now open the writer on "snapshot0" - make sure it succeeds
- writer = new IndexWriter(dir, getConfig(random, sdp).setIndexCommit(sdp.getSnapshot("snapshot0")));
+ writer = new IndexWriter(dir, getConfig(random(), sdp).setIndexCommit(sdp.getSnapshot("snapshot0")));
// this does the actual rollback
writer.commit();
writer.deleteUnusedFiles();
@@ -336,7 +336,7 @@ public class TestSnapshotDeletionPolicy
public void testReleaseSnapshot() throws Exception {
Directory dir = newDirectory();
SnapshotDeletionPolicy sdp = getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
prepareIndexAndSnapshots(sdp, writer, 1, "snapshot");
// Create another commit - we must do that, because otherwise the "snapshot"
@@ -368,13 +368,13 @@ public class TestSnapshotDeletionPolicy
int numSnapshots = 3;
Directory dir = newDirectory();
SnapshotDeletionPolicy sdp = getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
prepareIndexAndSnapshots(sdp, writer, numSnapshots, "snapshot");
writer.close();
// Make a new policy and initialize with snapshots.
sdp = getDeletionPolicy(sdp.getSnapshots());
- writer = new IndexWriter(dir, getConfig(random, sdp));
+ writer = new IndexWriter(dir, getConfig(random(), sdp));
// attempt to delete unused files - the snapshotted files should not be deleted
writer.deleteUnusedFiles();
writer.close();
@@ -386,7 +386,7 @@ public class TestSnapshotDeletionPolicy
public void testSnapshotLastCommitTwice() throws Exception {
Directory dir = newDirectory();
SnapshotDeletionPolicy sdp = getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
writer.addDocument(new Document());
writer.commit();
@@ -415,7 +415,7 @@ public class TestSnapshotDeletionPolicy
// on onInit().
Directory dir = newDirectory();
SnapshotDeletionPolicy sdp = getDeletionPolicy();
- IndexWriter writer = new IndexWriter(dir, getConfig(random, sdp));
+ IndexWriter writer = new IndexWriter(dir, getConfig(random(), sdp));
writer.addDocument(new Document());
writer.commit();
IndexCommit ic = sdp.snapshot("s1");
@@ -426,14 +426,14 @@ public class TestSnapshotDeletionPolicy
// open a new writer w/ KeepOnlyLastCommit policy, so it will delete "s1"
// commit.
- new IndexWriter(dir, getConfig(random, null)).close();
+ new IndexWriter(dir, getConfig(random(), null)).close();
assertFalse("snapshotted commit should not exist", dir.fileExists(ic.getSegmentsFileName()));
// Now reinit SDP from the commits in the index - the snapshot id should not
// exist anymore.
sdp = getDeletionPolicy(sdp.getSnapshots());
- new IndexWriter(dir, getConfig(random, sdp)).close();
+ new IndexWriter(dir, getConfig(random(), sdp)).close();
try {
sdp.getSnapshot("s1");
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java Sun Apr 15 14:41:44 2012
@@ -34,7 +34,7 @@ public class TestStressAdvance extends L
System.out.println("\nTEST: iter=" + iter);
}
Directory dir = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, dir);
+ RandomIndexWriter w = new RandomIndexWriter(random(), dir);
final Set<Integer> aDocs = new HashSet<Integer>();
final Document doc = new Document();
final Field f = newField("field", "", StringField.TYPE_UNSTORED);
@@ -43,7 +43,7 @@ public class TestStressAdvance extends L
doc.add(idField);
int num = atLeast(4097);
for(int id=0;id<num;id++) {
- if (random.nextInt(4) == 3) {
+ if (random().nextInt(4) == 3) {
f.setStringValue("a");
aDocs.add(id);
} else {
@@ -76,11 +76,11 @@ public class TestStressAdvance extends L
System.out.println("\nTEST: iter=" + iter + " iter2=" + iter2);
}
assertEquals(TermsEnum.SeekStatus.FOUND, te.seekCeil(new BytesRef("a")));
- de = _TestUtil.docs(random, te, null, de, false);
+ de = _TestUtil.docs(random(), te, null, de, false);
testOne(de, aDocIDs);
assertEquals(TermsEnum.SeekStatus.FOUND, te.seekCeil(new BytesRef("b")));
- de = _TestUtil.docs(random, te, null, de, false);
+ de = _TestUtil.docs(random(), te, null, de, false);
testOne(de, bDocIDs);
}
@@ -100,7 +100,7 @@ public class TestStressAdvance extends L
System.out.println(" cycle upto=" + upto + " of " + expected.size());
}
final int docID;
- if (random.nextInt(4) == 1 || upto == expected.size()-1) {
+ if (random().nextInt(4) == 1 || upto == expected.size()-1) {
// test nextDoc()
if (VERBOSE) {
System.out.println(" do nextDoc");
@@ -109,7 +109,7 @@ public class TestStressAdvance extends L
docID = docs.nextDoc();
} else {
// test advance()
- final int inc = _TestUtil.nextInt(random, 1, expected.size()-1-upto);
+ final int inc = _TestUtil.nextInt(random(), 1, expected.size()-1-upto);
if (VERBOSE) {
System.out.println(" do advance inc=" + inc);
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing.java Sun Apr 15 14:41:44 2012
@@ -77,7 +77,7 @@ public class TestStressIndexing extends
// Add 10 docs:
for(int j=0; j<10; j++) {
Document d = new Document();
- int n = random.nextInt();
+ int n = random().nextInt();
d.add(newField("id", Integer.toString(nextID++), StringField.TYPE_STORED));
d.add(newField("contents", English.intToEnglish(n), TextField.TYPE_UNSTORED));
writer.addDocument(d);
@@ -117,7 +117,7 @@ public class TestStressIndexing extends
*/
public void runStressTest(Directory directory, MergeScheduler mergeScheduler) throws Exception {
IndexWriter modifier = new IndexWriter(directory, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setOpenMode(OpenMode.CREATE).setMaxBufferedDocs(10).setMergeScheduler(
mergeScheduler));
modifier.commit();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java Sun Apr 15 14:41:44 2012
@@ -56,7 +56,7 @@ public class TestStressIndexing2 extends
@Override
boolean testPoint(String name) {
// if (name.equals("startCommit")) {
- if (random.nextInt(4) == 2)
+ if (random().nextInt(4) == 2)
Thread.yield();
return true;
}
@@ -69,7 +69,7 @@ public class TestStressIndexing2 extends
DocsAndWriter dw = indexRandomIWReader(5, 3, 100, dir);
DirectoryReader reader = dw.writer.getReader();
dw.writer.commit();
- verifyEquals(random, reader, dir, "id");
+ verifyEquals(random(), reader, dir, "id");
reader.close();
dw.writer.close();
dir.close();
@@ -79,10 +79,10 @@ public class TestStressIndexing2 extends
Directory dir1 = newDirectory();
Directory dir2 = newDirectory();
// mergeFactor=2; maxBufferedDocs=2; Map docs = indexRandom(1, 3, 2, dir1);
- int maxThreadStates = 1+random.nextInt(10);
- boolean doReaderPooling = random.nextBoolean();
+ int maxThreadStates = 1+random().nextInt(10);
+ boolean doReaderPooling = random().nextBoolean();
Map<String,Document> docs = indexRandom(5, 3, 100, dir1, maxThreadStates, doReaderPooling);
- indexSerial(random, docs, dir2);
+ indexSerial(random(), docs, dir2);
// verifying verify
// verifyEquals(dir1, dir1, "id");
@@ -101,16 +101,16 @@ public class TestStressIndexing2 extends
if (VERBOSE) {
System.out.println("\n\nTEST: top iter=" + i);
}
- sameFieldOrder=random.nextBoolean();
- mergeFactor=random.nextInt(3)+2;
- maxBufferedDocs=random.nextInt(3)+2;
- int maxThreadStates = 1+random.nextInt(10);
- boolean doReaderPooling = random.nextBoolean();
+ sameFieldOrder=random().nextBoolean();
+ mergeFactor=random().nextInt(3)+2;
+ maxBufferedDocs=random().nextInt(3)+2;
+ int maxThreadStates = 1+random().nextInt(10);
+ boolean doReaderPooling = random().nextBoolean();
seed++;
- int nThreads=random.nextInt(5)+1;
- int iter=random.nextInt(5)+1;
- int range=random.nextInt(20)+1;
+ int nThreads=random().nextInt(5)+1;
+ int iter=random().nextInt(5)+1;
+ int range=random().nextInt(20)+1;
Directory dir1 = newDirectory();
Directory dir2 = newDirectory();
if (VERBOSE) {
@@ -120,7 +120,7 @@ public class TestStressIndexing2 extends
if (VERBOSE) {
System.out.println("TEST: index serial");
}
- indexSerial(random, docs, dir2);
+ indexSerial(random(), docs, dir2);
if (VERBOSE) {
System.out.println("TEST: verify");
}
@@ -151,7 +151,7 @@ public class TestStressIndexing2 extends
public DocsAndWriter indexRandomIWReader(int nThreads, int iterations, int range, Directory dir) throws IOException, InterruptedException {
Map<String,Document> docs = new HashMap<String,Document>();
IndexWriter w = new MockIndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE).setRAMBufferSizeMB(
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE).setRAMBufferSizeMB(
0.1).setMaxBufferedDocs(maxBufferedDocs).setMergePolicy(newLogMergePolicy()));
w.commit();
LogMergePolicy lmp = (LogMergePolicy) w.getConfig().getMergePolicy();
@@ -202,7 +202,7 @@ public class TestStressIndexing2 extends
boolean doReaderPooling) throws IOException, InterruptedException {
Map<String,Document> docs = new HashMap<String,Document>();
IndexWriter w = new MockIndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.CREATE)
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.CREATE)
.setRAMBufferSizeMB(0.1).setMaxBufferedDocs(maxBufferedDocs).setIndexerThreadPool(new ThreadAffinityDocumentsWriterThreadPool(maxThreadStates))
.setReaderPooling(doReaderPooling).setMergePolicy(newLogMergePolicy()));
LogMergePolicy lmp = (LogMergePolicy) w.getConfig().getMergePolicy();
@@ -266,13 +266,13 @@ public class TestStressIndexing2 extends
w.close();
}
- public static void verifyEquals(Random r, DirectoryReader r1, Directory dir2, String idField) throws Throwable {
+ public void verifyEquals(Random r, DirectoryReader r1, Directory dir2, String idField) throws Throwable {
DirectoryReader r2 = IndexReader.open(dir2);
verifyEquals(r1, r2, idField);
r2.close();
}
- public static void verifyEquals(Directory dir1, Directory dir2, String idField) throws Throwable {
+ public void verifyEquals(Directory dir1, Directory dir2, String idField) throws Throwable {
DirectoryReader r1 = IndexReader.open(dir1);
DirectoryReader r2 = IndexReader.open(dir2);
verifyEquals(r1, r2, idField);
@@ -298,7 +298,7 @@ public class TestStressIndexing2 extends
}
- public static void verifyEquals(DirectoryReader r1, DirectoryReader r2, String idField) throws Throwable {
+ public void verifyEquals(DirectoryReader r1, DirectoryReader r2, String idField) throws Throwable {
if (VERBOSE) {
System.out.println("\nr1 docs:");
printDocs(r1);
@@ -336,7 +336,7 @@ public class TestStressIndexing2 extends
// deleted docs):
DocsEnum docs = null;
while(termsEnum.next() != null) {
- docs = _TestUtil.docs(random, termsEnum, null, docs, false);
+ docs = _TestUtil.docs(random(), termsEnum, null, docs, false);
while(docs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) {
fail("r1 is not empty but r2 is");
}
@@ -356,9 +356,9 @@ public class TestStressIndexing2 extends
break;
}
- termDocs1 = _TestUtil.docs(random, termsEnum, liveDocs1, termDocs1, false);
+ termDocs1 = _TestUtil.docs(random(), termsEnum, liveDocs1, termDocs1, false);
if (termsEnum2.seekExact(term, false)) {
- termDocs2 = _TestUtil.docs(random, termsEnum2, liveDocs2, termDocs2, false);
+ termDocs2 = _TestUtil.docs(random(), termsEnum2, liveDocs2, termDocs2, false);
} else {
termDocs2 = null;
}
@@ -417,7 +417,7 @@ public class TestStressIndexing2 extends
System.out.println(" pos=" + dpEnum.nextPosition());
}
} else {
- dEnum = _TestUtil.docs(random, termsEnum3, null, dEnum, true);
+ dEnum = _TestUtil.docs(random(), termsEnum3, null, dEnum, true);
assertNotNull(dEnum);
assertTrue(dEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS);
final int freq = dEnum.freq();
@@ -451,7 +451,7 @@ public class TestStressIndexing2 extends
System.out.println(" pos=" + dpEnum.nextPosition());
}
} else {
- dEnum = _TestUtil.docs(random, termsEnum3, null, dEnum, true);
+ dEnum = _TestUtil.docs(random(), termsEnum3, null, dEnum, true);
assertNotNull(dEnum);
assertTrue(dEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS);
final int freq = dEnum.freq();
@@ -508,7 +508,7 @@ public class TestStressIndexing2 extends
}
//System.out.println("TEST: term1=" + term1);
- docs1 = _TestUtil.docs(random, termsEnum1, liveDocs1, docs1, true);
+ docs1 = _TestUtil.docs(random(), termsEnum1, liveDocs1, docs1, true);
while (docs1.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) {
int d = docs1.docID();
int f = docs1.freq();
@@ -542,7 +542,7 @@ public class TestStressIndexing2 extends
}
//System.out.println("TEST: term1=" + term1);
- docs2 = _TestUtil.docs(random, termsEnum2, liveDocs2, docs2, true);
+ docs2 = _TestUtil.docs(random(), termsEnum2, liveDocs2, docs2, true);
while (docs2.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) {
int d = r2r1[docs2.docID()];
int f = docs2.freq();
@@ -669,8 +669,8 @@ public class TestStressIndexing2 extends
assertEquals(DocIdSetIterator.NO_MORE_DOCS, dpEnum1.nextDoc());
assertEquals(DocIdSetIterator.NO_MORE_DOCS, dpEnum2.nextDoc());
} else {
- dEnum1 = _TestUtil.docs(random, termsEnum1, null, dEnum1, true);
- dEnum2 = _TestUtil.docs(random, termsEnum2, null, dEnum2, true);
+ dEnum1 = _TestUtil.docs(random(), termsEnum1, null, dEnum1, true);
+ dEnum2 = _TestUtil.docs(random(), termsEnum2, null, dEnum2, true);
assertNotNull(dEnum1);
assertNotNull(dEnum2);
int docID1 = dEnum1.nextDoc();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressNRT.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressNRT.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressNRT.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestStressNRT.java Sun Apr 15 14:41:44 2012
@@ -66,20 +66,20 @@ public class TestStressNRT extends Lucen
public void test() throws Exception {
// update variables
- final int commitPercent = random.nextInt(20);
- final int softCommitPercent = random.nextInt(100); // what percent of the commits are soft
- final int deletePercent = random.nextInt(50);
- final int deleteByQueryPercent = random.nextInt(25);
+ final int commitPercent = random().nextInt(20);
+ final int softCommitPercent = random().nextInt(100); // what percent of the commits are soft
+ final int deletePercent = random().nextInt(50);
+ final int deleteByQueryPercent = random().nextInt(25);
final int ndocs = atLeast(50);
- final int nWriteThreads = _TestUtil.nextInt(random, 1, TEST_NIGHTLY ? 10 : 5);
- final int maxConcurrentCommits = _TestUtil.nextInt(random, 1, TEST_NIGHTLY ? 10 : 5); // number of committers at a time... needed if we want to avoid commit errors due to exceeding the max
+ final int nWriteThreads = _TestUtil.nextInt(random(), 1, TEST_NIGHTLY ? 10 : 5);
+ final int maxConcurrentCommits = _TestUtil.nextInt(random(), 1, TEST_NIGHTLY ? 10 : 5); // number of committers at a time... needed if we want to avoid commit errors due to exceeding the max
- final boolean tombstones = random.nextBoolean();
+ final boolean tombstones = random().nextBoolean();
// query variables
final AtomicLong operations = new AtomicLong(atLeast(10000)); // number of query operations to perform in total
- final int nReadThreads = _TestUtil.nextInt(random, 1, TEST_NIGHTLY ? 10 : 5);
+ final int nReadThreads = _TestUtil.nextInt(random(), 1, TEST_NIGHTLY ? 10 : 5);
initModel(ndocs);
final FieldType storedOnlyType = new FieldType();
@@ -106,14 +106,14 @@ public class TestStressNRT extends Lucen
Directory dir = newDirectory();
- final RandomIndexWriter writer = new RandomIndexWriter(random, dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ final RandomIndexWriter writer = new RandomIndexWriter(random(), dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
writer.setDoRandomForceMergeAssert(false);
writer.commit();
reader = IndexReader.open(dir);
for (int i=0; i<nWriteThreads; i++) {
Thread thread = new Thread("WRITER"+i) {
- Random rand = new Random(random.nextInt());
+ Random rand = new Random(random().nextInt());
@Override
public void run() {
@@ -137,7 +137,7 @@ public class TestStressNRT extends Lucen
DirectoryReader newReader;
if (rand.nextInt(100) < softCommitPercent) {
// assertU(h.commit("softCommit","true"));
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
if (VERBOSE) {
System.out.println("TEST: " + Thread.currentThread().getName() + ": call writer.getReader");
}
@@ -215,7 +215,7 @@ public class TestStressNRT extends Lucen
// set the lastId before we actually change it sometimes to try and
// uncover more race conditions between writing and reading
- boolean before = random.nextBoolean();
+ boolean before = random().nextBoolean();
if (before) {
lastId = id;
}
@@ -293,7 +293,7 @@ public class TestStressNRT extends Lucen
for (int i=0; i<nReadThreads; i++) {
Thread thread = new Thread("READER"+i) {
- Random rand = new Random(random.nextInt());
+ Random rand = new Random(random().nextInt());
@Override
public void run() {
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java Sun Apr 15 14:41:44 2012
@@ -35,7 +35,7 @@ public class TestSumDocFreq extends Luce
final int numDocs = atLeast(500);
Directory dir = newDirectory();
- RandomIndexWriter writer = new RandomIndexWriter(random, dir);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir);
Document doc = new Document();
Field id = newField("id", "", StringField.TYPE_UNSTORED);
@@ -46,11 +46,11 @@ public class TestSumDocFreq extends Luce
doc.add(field2);
for (int i = 0; i < numDocs; i++) {
id.setStringValue("" + i);
- char ch1 = (char) _TestUtil.nextInt(random, 'a', 'z');
- char ch2 = (char) _TestUtil.nextInt(random, 'a', 'z');
+ char ch1 = (char) _TestUtil.nextInt(random(), 'a', 'z');
+ char ch2 = (char) _TestUtil.nextInt(random(), 'a', 'z');
field1.setStringValue("" + ch1 + " " + ch2);
- ch1 = (char) _TestUtil.nextInt(random, 'a', 'z');
- ch2 = (char) _TestUtil.nextInt(random, 'a', 'z');
+ ch1 = (char) _TestUtil.nextInt(random(), 'a', 'z');
+ ch2 = (char) _TestUtil.nextInt(random(), 'a', 'z');
field2.setStringValue("" + ch1 + " " + ch2);
writer.addDocument(doc);
}
@@ -62,7 +62,7 @@ public class TestSumDocFreq extends Luce
int numDeletions = atLeast(20);
for (int i = 0; i < numDeletions; i++) {
- writer.deleteDocuments(new Term("id", "" + random.nextInt(numDocs)));
+ writer.deleteDocuments(new Term("id", "" + random().nextInt(numDocs)));
}
writer.forceMerge(1);
writer.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestTermVectorsReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestTermVectorsReader.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestTermVectorsReader.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestTermVectorsReader.java Sun Apr 15 14:41:44 2012
@@ -200,7 +200,7 @@ public class TestTermVectorsReader exten
}
public void testReader() throws IOException {
- TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random));
+ TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random()));
for (int j = 0; j < 5; j++) {
Terms vector = reader.get(j).terms(testFields[0]);
assertNotNull(vector);
@@ -219,7 +219,7 @@ public class TestTermVectorsReader exten
}
public void testDocsEnum() throws IOException {
- TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random));
+ TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random()));
for (int j = 0; j < 5; j++) {
Terms vector = reader.get(j).terms(testFields[0]);
assertNotNull(vector);
@@ -233,7 +233,7 @@ public class TestTermVectorsReader exten
//System.out.println("Term: " + term);
assertEquals(testTerms[i], term);
- docsEnum = _TestUtil.docs(random, termsEnum, null, docsEnum, false);
+ docsEnum = _TestUtil.docs(random(), termsEnum, null, docsEnum, false);
assertNotNull(docsEnum);
int doc = docsEnum.docID();
assertTrue(doc == -1 || doc == DocIdSetIterator.NO_MORE_DOCS);
@@ -246,7 +246,7 @@ public class TestTermVectorsReader exten
}
public void testPositionReader() throws IOException {
- TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random));
+ TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random()));
BytesRef[] terms;
Terms vector = reader.get(0).terms(testFields[0]);
assertNotNull(vector);
@@ -301,7 +301,7 @@ public class TestTermVectorsReader exten
}
public void testOffsetReader() throws IOException {
- TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random));
+ TermVectorsReader reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random()));
Terms vector = reader.get(0).terms(testFields[0]);
assertNotNull(vector);
TermsEnum termsEnum = vector.iterator(null);
@@ -343,7 +343,7 @@ public class TestTermVectorsReader exten
public void testBadParams() throws IOException {
TermVectorsReader reader = null;
try {
- reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random));
+ reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random()));
//Bad document number, good field number
reader.get(50);
fail();
@@ -352,7 +352,7 @@ public class TestTermVectorsReader exten
} finally {
reader.close();
}
- reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random));
+ reader = Codec.getDefault().termVectorsFormat().vectorsReader(dir, seg, fieldInfos, newIOContext(random()));
//good document number, bad field
Terms vector = reader.get(0).terms("f50");
assertNull(vector);