You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by dw...@apache.org on 2012/04/15 16:42:01 UTC
svn commit: r1326351 [7/22] - in /lucene/dev/trunk: ./ dev-tools/eclipse/
lucene/
lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/
lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/custom/
lucene/contrib/high...
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java Sun Apr 15 14:41:44 2012
@@ -34,7 +34,7 @@ public class TestIndexWriterMergePolicy
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(10).setMergePolicy(new LogDocMergePolicy()));
for (int i = 0; i < 100; i++) {
@@ -51,7 +51,7 @@ public class TestIndexWriterMergePolicy
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(10).setMergePolicy(new LogDocMergePolicy()));
boolean noOverMerge = false;
@@ -76,7 +76,7 @@ public class TestIndexWriterMergePolicy
mp.setMinMergeDocs(100);
mp.setMergeFactor(10);
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(10).setMergePolicy(mp));
for (int i = 0; i < 100; i++) {
@@ -86,7 +86,7 @@ public class TestIndexWriterMergePolicy
mp = new LogDocMergePolicy();
mp.setMergeFactor(10);
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(
+ new MockAnalyzer(random())).setOpenMode(
OpenMode.APPEND).setMaxBufferedDocs(10).setMergePolicy(mp));
mp.setMinMergeDocs(100);
checkInvariants(writer);
@@ -102,7 +102,7 @@ public class TestIndexWriterMergePolicy
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(10).
setMergePolicy(newLogMergePolicy()).
setMergeScheduler(new SerialMergeScheduler())
@@ -131,7 +131,7 @@ public class TestIndexWriterMergePolicy
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(101).setMergePolicy(new LogDocMergePolicy())
.setMergeScheduler(new SerialMergeScheduler()));
@@ -145,7 +145,7 @@ public class TestIndexWriterMergePolicy
writer.close();
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(
+ new MockAnalyzer(random())).setOpenMode(
OpenMode.APPEND).setMaxBufferedDocs(101).setMergePolicy(new LogDocMergePolicy())
.setMergeScheduler(new SerialMergeScheduler()));
}
@@ -154,7 +154,7 @@ public class TestIndexWriterMergePolicy
LogDocMergePolicy ldmp = new LogDocMergePolicy();
ldmp.setMergeFactor(10);
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(
+ new MockAnalyzer(random())).setOpenMode(
OpenMode.APPEND).setMaxBufferedDocs(10).setMergePolicy(ldmp).setMergeScheduler(new SerialMergeScheduler()));
// merge policy only fixes segments on levels where merges
@@ -183,7 +183,7 @@ public class TestIndexWriterMergePolicy
LogDocMergePolicy ldmp = new LogDocMergePolicy();
ldmp.setMergeFactor(100);
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(10).setMergePolicy(ldmp));
for (int i = 0; i < 250; i++) {
@@ -195,7 +195,7 @@ public class TestIndexWriterMergePolicy
// delete some docs without merging
writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergePolicy(NoMergePolicy.NO_COMPOUND_FILES)
);
writer.deleteDocuments(new Term("content", "aaa"));
@@ -204,7 +204,7 @@ public class TestIndexWriterMergePolicy
ldmp = new LogDocMergePolicy();
ldmp.setMergeFactor(5);
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setOpenMode(
+ new MockAnalyzer(random())).setOpenMode(
OpenMode.APPEND).setMaxBufferedDocs(10).setMergePolicy(ldmp).setMergeScheduler(new ConcurrentMergeScheduler()));
// merge factor is changed, so check invariants after all adds
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMerging.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMerging.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMerging.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterMerging.java Sun Apr 15 14:41:44 2012
@@ -44,14 +44,14 @@ public class TestIndexWriterMerging exte
Directory indexA = newDirectory();
Directory indexB = newDirectory();
- fillIndex(random, indexA, 0, num);
+ fillIndex(random(), indexA, 0, num);
boolean fail = verifyIndex(indexA, 0);
if (fail)
{
fail("Index a is invalid");
}
- fillIndex(random, indexB, num, num);
+ fillIndex(random(), indexB, num, num);
fail = verifyIndex(indexB, num);
if (fail)
{
@@ -62,7 +62,7 @@ public class TestIndexWriterMerging exte
IndexWriter writer = new IndexWriter(
merged,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergePolicy(newLogMergePolicy(2))
);
writer.addIndexes(indexA, indexB);
@@ -123,7 +123,7 @@ public class TestIndexWriterMerging exte
public void testForceMergeDeletes() throws IOException {
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(2).setRAMBufferSizeMB(
IndexWriterConfig.DISABLE_AUTO_FLUSH));
Document document = new Document();
@@ -154,7 +154,7 @@ public class TestIndexWriterMerging exte
assertEquals(10, ir.numDocs());
ir.close();
- IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
writer = new IndexWriter(dir, dontMergeConfig);
writer.deleteDocuments(new Term("id", "0"));
@@ -165,7 +165,7 @@ public class TestIndexWriterMerging exte
assertEquals(8, ir.numDocs());
ir.close();
- writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy()));
+ writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(newLogMergePolicy()));
assertEquals(8, writer.numDocs());
assertEquals(10, writer.maxDoc());
writer.forceMergeDeletes();
@@ -183,7 +183,7 @@ public class TestIndexWriterMerging exte
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(2).
setRAMBufferSizeMB(IndexWriterConfig.DISABLE_AUTO_FLUSH).
setMergePolicy(newLogMergePolicy(50))
@@ -217,7 +217,7 @@ public class TestIndexWriterMerging exte
assertEquals(98, ir.numDocs());
ir.close();
- IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
writer = new IndexWriter(dir, dontMergeConfig);
for(int i=0;i<98;i+=2) {
@@ -231,7 +231,7 @@ public class TestIndexWriterMerging exte
writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergePolicy(newLogMergePolicy(3))
);
assertEquals(49, writer.numDocs());
@@ -250,7 +250,7 @@ public class TestIndexWriterMerging exte
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(2).
setRAMBufferSizeMB(IndexWriterConfig.DISABLE_AUTO_FLUSH).
setMergePolicy(newLogMergePolicy(50))
@@ -283,7 +283,7 @@ public class TestIndexWriterMerging exte
assertEquals(98, ir.numDocs());
ir.close();
- IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriterConfig dontMergeConfig = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
writer = new IndexWriter(dir, dontMergeConfig);
for(int i=0;i<98;i+=2) {
@@ -296,7 +296,7 @@ public class TestIndexWriterMerging exte
writer = new IndexWriter(
dir,
- newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergePolicy(newLogMergePolicy(3))
);
writer.forceMergeDeletes(false);
@@ -335,7 +335,7 @@ public class TestIndexWriterMerging exte
public void testSetMaxMergeDocs() throws IOException {
Directory dir = newDirectory();
IndexWriterConfig conf = newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMergeScheduler(new MyMergeScheduler()).setMaxBufferedDocs(2).setMergePolicy(newLogMergePolicy());
LogMergePolicy lmp = (LogMergePolicy) conf.getMergePolicy();
lmp.setMaxMergeDocs(20);
@@ -369,7 +369,7 @@ public class TestIndexWriterMerging exte
}
IndexWriterConfig conf = newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setOpenMode(OpenMode.CREATE).
setMaxBufferedDocs(2).
setMergePolicy(newLogMergePolicy());
@@ -442,7 +442,7 @@ public class TestIndexWriterMerging exte
reader.close();
// Reopen
- writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND).setMergePolicy(newLogMergePolicy()));
+ writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND).setMergePolicy(newLogMergePolicy()));
}
writer.close();
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java Sun Apr 15 14:41:44 2012
@@ -40,13 +40,13 @@ public class TestIndexWriterNRTIsCurrent
LockObtainFailedException, IOException, InterruptedException {
Directory dir = newDirectory();
IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random));
+ new MockAnalyzer(random()));
IndexWriter writer = new IndexWriter(dir, conf);
ReaderHolder holder = new ReaderHolder();
ReaderThread[] threads = new ReaderThread[atLeast(3)];
final CountDownLatch latch = new CountDownLatch(1);
WriterThread writerThread = new WriterThread(holder, writer,
- atLeast(500), random, latch);
+ atLeast(500), random(), latch);
for (int i = 0; i < threads.length; i++) {
threads[i] = new ReaderThread(holder, latch);
threads[i].start();
@@ -74,7 +74,6 @@ public class TestIndexWriterNRTIsCurrent
private final ReaderHolder holder;
private final IndexWriter writer;
private final int numOps;
- private final Random random;
private boolean countdown = true;
private final CountDownLatch latch;
Throwable failed;
@@ -85,12 +84,12 @@ public class TestIndexWriterNRTIsCurrent
this.holder = holder;
this.writer = writer;
this.numOps = numOps;
- this.random = random;
this.latch = latch;
}
public void run() {
DirectoryReader currentReader = null;
+ Random random = LuceneTestCase.random();
try {
Document doc = new Document();
doc.add(new Field("id", "1", TextField.TYPE_UNSTORED));
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java Sun Apr 15 14:41:44 2012
@@ -53,14 +53,14 @@ public class TestIndexWriterOnDiskFull e
System.out.println("TEST: pass=" + pass);
}
boolean doAbort = pass == 1;
- long diskFree = _TestUtil.nextInt(random, 100, 300);
+ long diskFree = _TestUtil.nextInt(random(), 100, 300);
while(true) {
if (VERBOSE) {
System.out.println("TEST: cycle: diskFree=" + diskFree);
}
- MockDirectoryWrapper dir = new MockDirectoryWrapper(random, new RAMDirectory());
+ MockDirectoryWrapper dir = new MockDirectoryWrapper(random(), new RAMDirectory());
dir.setMaxSizeInBytes(diskFree);
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
MergeScheduler ms = writer.getConfig().getMergeScheduler();
if (ms instanceof ConcurrentMergeScheduler) {
// This test intentionally produces exceptions
@@ -120,7 +120,7 @@ public class TestIndexWriterOnDiskFull e
dir.close();
// Now try again w/ more space:
- diskFree += TEST_NIGHTLY ? _TestUtil.nextInt(random, 400, 600) : _TestUtil.nextInt(random, 3000, 5000);
+ diskFree += TEST_NIGHTLY ? _TestUtil.nextInt(random(), 400, 600) : _TestUtil.nextInt(random(), 3000, 5000);
} else {
//_TestUtil.syncConcurrentMerges(writer);
dir.setMaxSizeInBytes(0);
@@ -168,7 +168,7 @@ public class TestIndexWriterOnDiskFull e
long inputDiskUsage = 0;
for(int i=0;i<NUM_DIR;i++) {
dirs[i] = newDirectory();
- IndexWriter writer = new IndexWriter(dirs[i], newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(dirs[i], newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
for(int j=0;j<25;j++) {
addDocWithIndex(writer, 25*i+j);
}
@@ -182,7 +182,7 @@ public class TestIndexWriterOnDiskFull e
// Now, build a starting index that has START_COUNT docs. We
// will then try to addIndexes into a copy of this:
MockDirectoryWrapper startDir = newDirectory();
- IndexWriter writer = new IndexWriter(startDir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(startDir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
for(int j=0;j<START_COUNT;j++) {
addDocWithIndex(writer, j);
}
@@ -224,7 +224,7 @@ public class TestIndexWriterOnDiskFull e
System.out.println("TEST: iter=" + iter);
// Start with 100 bytes more than we are currently using:
- long diskFree = diskUsage+_TestUtil.nextInt(random, 50, 200);
+ long diskFree = diskUsage+_TestUtil.nextInt(random(), 50, 200);
int method = iter;
@@ -246,8 +246,8 @@ public class TestIndexWriterOnDiskFull e
}
// Make a new dir that will enforce disk usage:
- MockDirectoryWrapper dir = new MockDirectoryWrapper(random, new RAMDirectory(startDir, newIOContext(random)));
- writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND).setMergePolicy(newLogMergePolicy()));
+ MockDirectoryWrapper dir = new MockDirectoryWrapper(random(), new RAMDirectory(startDir, newIOContext(random())));
+ writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setOpenMode(OpenMode.APPEND).setMergePolicy(newLogMergePolicy()));
IOException err = null;
MergeScheduler ms = writer.getConfig().getMergeScheduler();
@@ -440,7 +440,7 @@ public class TestIndexWriterOnDiskFull e
dir.close();
// Try again with more free space:
- diskFree += TEST_NIGHTLY ? _TestUtil.nextInt(random, 4000, 8000) : _TestUtil.nextInt(random, 40000, 80000);
+ diskFree += TEST_NIGHTLY ? _TestUtil.nextInt(random(), 4000, 8000) : _TestUtil.nextInt(random(), 40000, 80000);
}
}
@@ -478,7 +478,7 @@ public class TestIndexWriterOnDiskFull e
//IndexWriter w = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setReaderPooling(true));
IndexWriter w = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergeScheduler(new SerialMergeScheduler()).
setReaderPooling(true).
setMergePolicy(newLogMergePolicy(2))
@@ -519,7 +519,7 @@ public class TestIndexWriterOnDiskFull e
// OK:
public void testImmediateDiskFull() throws IOException {
MockDirectoryWrapper dir = newDirectory();
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(2).setMergeScheduler(new ConcurrentMergeScheduler()));
dir.setMaxSizeInBytes(Math.max(1, dir.getRecomputedActualSizeInBytes()));
final Document doc = new Document();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnJRECrash.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnJRECrash.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnJRECrash.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterOnJRECrash.java Sun Apr 15 14:41:44 2012
@@ -64,7 +64,7 @@ public class TestIndexWriterOnJRECrash e
}
} else {
// we are the fork, setup a crashing thread
- final int crashTime = _TestUtil.nextInt(random, 3000, 4000);
+ final int crashTime = _TestUtil.nextInt(random(), 3000, 4000);
Thread t = new Thread() {
@Override
public void run() {
@@ -96,7 +96,7 @@ public class TestIndexWriterOnJRECrash e
// passing NIGHTLY to this test makes it run for much longer, easier to catch it in the act...
cmd.add("-Dtests.nightly=true");
cmd.add("-DtempDir=" + tempDir.getPath());
- cmd.add("-Dtests.seed=" + random.nextLong() + ":" + random.nextLong());
+ cmd.add("-Dtests.seed=" + random().nextLong() + ":" + random().nextLong());
cmd.add("-ea");
cmd.add("-cp");
cmd.add(System.getProperty("java.class.path"));
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java Sun Apr 15 14:41:44 2012
@@ -51,7 +51,7 @@ public class TestIndexWriterReader exten
public static int count(Term t, IndexReader r) throws IOException {
int count = 0;
- DocsEnum td = _TestUtil.docs(random, r,
+ DocsEnum td = _TestUtil.docs(random(), r,
t.field(), new BytesRef(t.text()),
MultiFields.getLiveDocs(r),
null,
@@ -68,26 +68,26 @@ public class TestIndexWriterReader exten
public void testAddCloseOpen() throws IOException {
Directory dir1 = newDirectory();
- IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
IndexWriter writer = new IndexWriter(dir1, iwc);
for (int i = 0; i < 97 ; i++) {
DirectoryReader reader = writer.getReader();
if (i == 0) {
- writer.addDocument(DocHelper.createDocument(i, "x", 1 + random.nextInt(5)));
+ writer.addDocument(DocHelper.createDocument(i, "x", 1 + random().nextInt(5)));
} else {
- int previous = random.nextInt(i);
+ int previous = random().nextInt(i);
// a check if the reader is current here could fail since there might be
// merges going on.
- switch (random.nextInt(5)) {
+ switch (random().nextInt(5)) {
case 0:
case 1:
case 2:
- writer.addDocument(DocHelper.createDocument(i, "x", 1 + random.nextInt(5)));
+ writer.addDocument(DocHelper.createDocument(i, "x", 1 + random().nextInt(5)));
break;
case 3:
writer.updateDocument(new Term("id", "" + previous), DocHelper.createDocument(
- previous, "x", 1 + random.nextInt(5)));
+ previous, "x", 1 + random().nextInt(5)));
break;
case 4:
writer.deleteDocuments(new Term("id", "" + previous));
@@ -102,10 +102,10 @@ public class TestIndexWriterReader exten
assertTrue(reader.isCurrent());
writer.close();
assertTrue(reader.isCurrent()); // all changes are visible to the reader
- iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
writer = new IndexWriter(dir1, iwc);
assertTrue(reader.isCurrent());
- writer.addDocument(DocHelper.createDocument(1, "x", 1+random.nextInt(5)));
+ writer.addDocument(DocHelper.createDocument(1, "x", 1+random().nextInt(5)));
assertTrue(reader.isCurrent()); // segments in ram but IW is different to the readers one
writer.close();
assertFalse(reader.isCurrent()); // segments written
@@ -117,12 +117,12 @@ public class TestIndexWriterReader exten
boolean doFullMerge = true;
Directory dir1 = newDirectory();
- IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
if (iwc.getMaxBufferedDocs() < 20) {
iwc.setMaxBufferedDocs(20);
}
// no merging
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
iwc.setMergePolicy(NoMergePolicy.NO_COMPOUND_FILES);
} else {
iwc.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
@@ -167,7 +167,7 @@ public class TestIndexWriterReader exten
assertEquals(0, count(new Term("id", id10), r3));
assertEquals(1, count(new Term("id", Integer.toString(8000)), r3));
- writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document doc = new Document();
doc.add(newField("field", "a b c", TextField.TYPE_UNSTORED));
writer.addDocument(doc);
@@ -187,7 +187,7 @@ public class TestIndexWriterReader exten
public void testIsCurrent() throws IOException {
Directory dir = newDirectory();
- IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
IndexWriter writer = new IndexWriter(dir, iwc);
Document doc = new Document();
@@ -195,7 +195,7 @@ public class TestIndexWriterReader exten
writer.addDocument(doc);
writer.close();
- iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
writer = new IndexWriter(dir, iwc);
doc = new Document();
doc.add(newField("field", "a b c", TextField.TYPE_UNSTORED));
@@ -232,12 +232,12 @@ public class TestIndexWriterReader exten
boolean doFullMerge = false;
Directory dir1 = newDirectory();
- IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
if (iwc.getMaxBufferedDocs() < 20) {
iwc.setMaxBufferedDocs(20);
}
// no merging
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
iwc.setMergePolicy(NoMergePolicy.NO_COMPOUND_FILES);
} else {
iwc.setMergePolicy(NoMergePolicy.COMPOUND_FILES);
@@ -250,7 +250,7 @@ public class TestIndexWriterReader exten
// create a 2nd index
Directory dir2 = newDirectory();
- IndexWriter writer2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
createIndexNoClose(!doFullMerge, "index2", writer2);
writer2.close();
@@ -287,11 +287,11 @@ public class TestIndexWriterReader exten
boolean doFullMerge = false;
Directory dir1 = newDirectory();
- IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
// create a 2nd index
Directory dir2 = newDirectory();
- IndexWriter writer2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
createIndexNoClose(!doFullMerge, "index2", writer2);
writer2.close();
@@ -319,7 +319,7 @@ public class TestIndexWriterReader exten
boolean doFullMerge = true;
Directory dir1 = newDirectory();
- IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setReaderTermsIndexDivisor(2));
+ IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setReaderTermsIndexDivisor(2));
// create the index
createIndexNoClose(!doFullMerge, "index1", writer);
writer.flush(false, true);
@@ -356,7 +356,7 @@ public class TestIndexWriterReader exten
writer.close();
// reopen the writer to verify the delete made it to the directory
- writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
IndexReader w2r1 = writer.getReader();
assertEquals(0, count(new Term("id", id10), w2r1));
w2r1.close();
@@ -369,7 +369,7 @@ public class TestIndexWriterReader exten
int numDirs = 3;
Directory mainDir = newDirectory();
- IndexWriter mainWriter = new IndexWriter(mainDir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy()));
+ IndexWriter mainWriter = new IndexWriter(mainDir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(newLogMergePolicy()));
_TestUtil.reduceOpenFiles(mainWriter);
AddDirectoriesThreads addDirThreads = new AddDirectoriesThreads(numIter, mainWriter);
@@ -412,7 +412,7 @@ public class TestIndexWriterReader exten
this.numDirs = numDirs;
this.mainWriter = mainWriter;
addDir = newDirectory();
- IndexWriter writer = new IndexWriter(addDir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2));
+ IndexWriter writer = new IndexWriter(addDir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMaxBufferedDocs(2));
for (int i = 0; i < NUM_INIT_DOCS; i++) {
Document doc = DocHelper.createDocument(i, "addindex", 4);
writer.addDocument(doc);
@@ -463,7 +463,7 @@ public class TestIndexWriterReader exten
try {
final Directory[] dirs = new Directory[numDirs];
for (int k = 0; k < numDirs; k++)
- dirs[k] = new MockDirectoryWrapper(random, new RAMDirectory(addDir, newIOContext(random)));
+ dirs[k] = new MockDirectoryWrapper(random(), new RAMDirectory(addDir, newIOContext(random())));
//int j = 0;
//while (true) {
// System.out.println(Thread.currentThread().getName() + ": iter
@@ -521,7 +521,7 @@ public class TestIndexWriterReader exten
*/
public void doTestIndexWriterReopenSegment(boolean doFullMerge) throws Exception {
Directory dir1 = newDirectory();
- IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
IndexReader r1 = writer.getReader();
assertEquals(0, r1.maxDoc());
createIndexNoClose(false, "index1", writer);
@@ -557,7 +557,7 @@ public class TestIndexWriterReader exten
writer.close();
// test whether the changes made it to the directory
- writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
IndexReader w2r1 = writer.getReader();
// insure the deletes were actually flushed to the directory
assertEquals(200, w2r1.maxDoc());
@@ -615,7 +615,7 @@ public class TestIndexWriterReader exten
MyWarmer warmer = new MyWarmer();
IndexWriter writer = new IndexWriter(
dir1,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(2).
setMergedSegmentWarmer(warmer).
setMergeScheduler(new ConcurrentMergeScheduler()).
@@ -650,7 +650,7 @@ public class TestIndexWriterReader exten
public void testAfterCommit() throws Exception {
Directory dir1 = newDirectory();
- IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergeScheduler(new ConcurrentMergeScheduler()));
+ IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergeScheduler(new ConcurrentMergeScheduler()));
writer.commit();
// create the index
@@ -682,7 +682,7 @@ public class TestIndexWriterReader exten
// Make sure reader remains usable even if IndexWriter closes
public void testAfterClose() throws Exception {
Directory dir1 = newDirectory();
- IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
// create the index
createIndexNoClose(false, "test", writer);
@@ -712,7 +712,7 @@ public class TestIndexWriterReader exten
MockDirectoryWrapper dir1 = newDirectory();
final IndexWriter writer = new IndexWriter(
dir1,
- newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergePolicy(newLogMergePolicy(2))
);
@@ -722,7 +722,7 @@ public class TestIndexWriterReader exten
final Directory[] dirs = new Directory[10];
for (int i=0;i<10;i++) {
- dirs[i] = new MockDirectoryWrapper(random, new RAMDirectory(dir1, newIOContext(random)));
+ dirs[i] = new MockDirectoryWrapper(random(), new RAMDirectory(dir1, newIOContext(random())));
}
DirectoryReader r = writer.getReader();
@@ -795,7 +795,7 @@ public class TestIndexWriterReader exten
Directory dir1 = newDirectory();
final IndexWriter writer = new IndexWriter(
dir1,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMergePolicy(newLogMergePolicy(2))
);
@@ -813,7 +813,7 @@ public class TestIndexWriterReader exten
final Thread[] threads = new Thread[numThreads];
for(int i=0;i<numThreads;i++) {
threads[i] = new Thread() {
- final Random r = new Random(random.nextLong());
+ final Random r = new Random(random().nextLong());
@Override
public void run() {
@@ -875,7 +875,7 @@ public class TestIndexWriterReader exten
public void testForceMergeDeletes() throws Throwable {
Directory dir = newDirectory();
- final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy()));
+ final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(newLogMergePolicy()));
Document doc = new Document();
doc.add(newField("field", "a b c", TextField.TYPE_UNSTORED));
Field id = newField("id", "", StringField.TYPE_UNSTORED);
@@ -899,7 +899,7 @@ public class TestIndexWriterReader exten
public void testDeletesNumDocs() throws Throwable {
Directory dir = newDirectory();
- final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ final IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document doc = new Document();
doc.add(newField("field", "a b c", TextField.TYPE_UNSTORED));
Field id = newField("id", "", StringField.TYPE_UNSTORED);
@@ -929,7 +929,7 @@ public class TestIndexWriterReader exten
public void testEmptyIndex() throws Exception {
// Ensures that getReader works on an empty index, which hasn't been committed yet.
Directory dir = newDirectory();
- IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
IndexReader r = w.getReader();
assertEquals(0, r.numDocs());
r.close();
@@ -942,7 +942,7 @@ public class TestIndexWriterReader exten
final AtomicBoolean didWarm = new AtomicBoolean();
IndexWriter w = new IndexWriter(
dir,
- newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(2).
setReaderPooling(true).
setMergedSegmentWarmer(new IndexWriter.IndexReaderWarmer() {
@@ -975,7 +975,7 @@ public class TestIndexWriterReader exten
"Lucene3x".equals(Codec.getDefault().getName()));
IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random)).setReaderTermsIndexDivisor(-1);
+ new MockAnalyzer(random())).setReaderTermsIndexDivisor(-1);
// Don't proceed if picked Codec is in the list of illegal ones.
final String format = _TestUtil.getPostingsFormat("f");
@@ -989,7 +989,7 @@ public class TestIndexWriterReader exten
w.addDocument(doc);
IndexReader r = IndexReader.open(w, true).getSequentialSubReaders()[0];
try {
- _TestUtil.docs(random, r, "f", new BytesRef("val"), null, null, false);
+ _TestUtil.docs(random(), r, "f", new BytesRef("val"), null, null, false);
fail("should have failed to seek since terms index was not loaded.");
} catch (IllegalStateException e) {
// expected - we didn't load the term index
@@ -1004,7 +1004,7 @@ public class TestIndexWriterReader exten
Directory d = newDirectory();
IndexWriter w = new IndexWriter(
d,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
DirectoryReader r = w.getReader(); // start pooling readers
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterUnicode.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterUnicode.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterUnicode.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterUnicode.java Sun Apr 15 14:41:44 2012
@@ -64,7 +64,7 @@ public class TestIndexWriterUnicode exte
};
private int nextInt(int lim) {
- return random.nextInt(lim);
+ return random().nextInt(lim);
}
private int nextInt(int start, int end) {
@@ -98,7 +98,7 @@ public class TestIndexWriterUnicode exte
else if (5 == t && i < len-1) {
// Illegal unpaired surrogate
if (nextInt(10) == 7) {
- if (random.nextBoolean())
+ if (random().nextBoolean())
buffer[i] = (char) nextInt(0xd800, 0xdc00);
else
buffer[i] = (char) nextInt(0xdc00, 0xe000);
@@ -235,7 +235,7 @@ public class TestIndexWriterUnicode exte
public void testEmbeddedFFFF() throws Throwable {
Directory d = newDirectory();
- IndexWriter w = new IndexWriter(d, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter w = new IndexWriter(d, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document doc = new Document();
doc.add(newField("field", "a a\uffffb", TextField.TYPE_UNSTORED));
w.addDocument(doc);
@@ -274,7 +274,7 @@ public class TestIndexWriterUnicode exte
// Make sure terms, including ones with surrogate pairs,
// sort in codepoint sort order by default
public void testTermUTF16SortOrder() throws Throwable {
- Random rnd = random;
+ Random rnd = random();
Directory dir = newDirectory();
RandomIndexWriter writer = new RandomIndexWriter(rnd, dir);
Document d = new Document();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java Sun Apr 15 14:41:44 2012
@@ -126,7 +126,7 @@ public class TestIndexWriterWithThreads
MockDirectoryWrapper dir = newDirectory();
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(2).
setMergeScheduler(new ConcurrentMergeScheduler()).
setMergePolicy(newLogMergePolicy(4))
@@ -170,7 +170,7 @@ public class TestIndexWriterWithThreads
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(10).
setMergeScheduler(new ConcurrentMergeScheduler()).
setMergePolicy(newLogMergePolicy(4))
@@ -211,7 +211,7 @@ public class TestIndexWriterWithThreads
// Quick test to make sure index is not corrupt:
IndexReader reader = IndexReader.open(dir);
- DocsEnum tdocs = _TestUtil.docs(random, reader,
+ DocsEnum tdocs = _TestUtil.docs(random(), reader,
"field",
new BytesRef("aaa"),
MultiFields.getLiveDocs(reader),
@@ -242,7 +242,7 @@ public class TestIndexWriterWithThreads
IndexWriter writer = new IndexWriter(
dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(2).
setMergeScheduler(new ConcurrentMergeScheduler()).
setMergePolicy(newLogMergePolicy(4))
@@ -297,7 +297,7 @@ public class TestIndexWriterWithThreads
public void _testSingleThreadFailure(MockDirectoryWrapper.Failure failure) throws IOException {
MockDirectoryWrapper dir = newDirectory();
- IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setMaxBufferedDocs(2).setMergeScheduler(new ConcurrentMergeScheduler()));
final Document doc = new Document();
FieldType customType = new FieldType(TextField.TYPE_STORED);
@@ -480,7 +480,7 @@ public class TestIndexWriterWithThreads
Field field = newField("field", "testData", TextField.TYPE_STORED);
doc.add(field);
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ TEST_VERSION_CURRENT, new MockAnalyzer(random())));
iwConstructed.countDown();
startIndexing.await();
writer.addDocument(doc);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java Sun Apr 15 14:41:44 2012
@@ -100,7 +100,7 @@ public class TestIndexableField extends
@Override
public float boost() {
- return 1.0f + random.nextFloat();
+ return 1.0f + random().nextFloat();
}
@Override
@@ -157,7 +157,7 @@ public class TestIndexableField extends
public void testArbitraryFields() throws Exception {
final Directory dir = newDirectory();
- final RandomIndexWriter w = new RandomIndexWriter(random, dir);
+ final RandomIndexWriter w = new RandomIndexWriter(random(), dir);
final int NUM_DOCS = atLeast(27);
if (VERBOSE) {
@@ -167,7 +167,7 @@ public class TestIndexableField extends
int baseCount = 0;
for(int docCount=0;docCount<NUM_DOCS;docCount++) {
- final int fieldCount = _TestUtil.nextInt(random, 1, 17);
+ final int fieldCount = _TestUtil.nextInt(random(), 1, 17);
fieldsPerDoc[docCount] = fieldCount-1;
final int finalDocCount = docCount;
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIsCurrent.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIsCurrent.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIsCurrent.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestIsCurrent.java Sun Apr 15 14:41:44 2012
@@ -38,7 +38,7 @@ public class TestIsCurrent extends Lucen
// initialize directory
directory = newDirectory();
- writer = new RandomIndexWriter(random, directory);
+ writer = new RandomIndexWriter(random(), directory);
// write document
Document doc = new Document();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLazyProxSkipping.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLazyProxSkipping.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLazyProxSkipping.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLazyProxSkipping.java Sun Apr 15 14:41:44 2012
@@ -50,7 +50,7 @@ public class TestLazyProxSkipping extend
private class SeekCountingDirectory extends MockDirectoryWrapper {
public SeekCountingDirectory(Directory delegate) {
- super(random, delegate);
+ super(random(), delegate);
}
@Override
@@ -143,7 +143,7 @@ public class TestLazyProxSkipping extend
public void testSeek() throws IOException {
Directory directory = newDirectory();
- IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
for (int i = 0; i < 10; i++) {
Document doc = new Document();
doc.add(newField(this.field, "a b", TextField.TYPE_STORED));
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java Sun Apr 15 14:41:44 2012
@@ -43,9 +43,9 @@ public class TestLongPostings extends Lu
// Produces a realistic unicode random string that
// survives MockAnalyzer unchanged:
private String getRandomTerm(String other) throws IOException {
- Analyzer a = new MockAnalyzer(random);
+ Analyzer a = new MockAnalyzer(random());
while(true) {
- String s = _TestUtil.randomRealisticUnicodeString(random);
+ String s = _TestUtil.randomRealisticUnicodeString(random());
if (other != null && s.equals(other)) {
continue;
}
@@ -80,7 +80,7 @@ public class TestLongPostings extends Lu
public void testLongPostings() throws Exception {
// Don't use _TestUtil.getTempDir so that we own the
// randomness (ie same seed will point to same dir):
- Directory dir = newFSDirectory(_TestUtil.getTempDir("longpostings" + "." + random.nextLong()));
+ Directory dir = newFSDirectory(_TestUtil.getTempDir("longpostings" + "." + random().nextLong()));
final int NUM_DOCS = atLeast(2000);
@@ -105,24 +105,24 @@ public class TestLongPostings extends Lu
final FixedBitSet isS1 = new FixedBitSet(NUM_DOCS);
for(int idx=0;idx<NUM_DOCS;idx++) {
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
isS1.set(idx);
}
}
final IndexReader r;
- final IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ final IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setOpenMode(IndexWriterConfig.OpenMode.CREATE)
.setMergePolicy(newLogMergePolicy());
- iwc.setRAMBufferSizeMB(16.0 + 16.0 * random.nextDouble());
+ iwc.setRAMBufferSizeMB(16.0 + 16.0 * random().nextDouble());
iwc.setMaxBufferedDocs(-1);
- final RandomIndexWriter riw = new RandomIndexWriter(random, dir, iwc);
+ final RandomIndexWriter riw = new RandomIndexWriter(random(), dir, iwc);
for(int idx=0;idx<NUM_DOCS;idx++) {
final Document doc = new Document();
String s = isS1.get(idx) ? s1 : s2;
final Field f = newField("field", s, TextField.TYPE_UNSTORED);
- final int count = _TestUtil.nextInt(random, 1, 4);
+ final int count = _TestUtil.nextInt(random(), 1, 4);
for(int ct=0;ct<count;ct++) {
doc.add(f);
}
@@ -158,7 +158,7 @@ public class TestLongPostings extends Lu
final String term;
final boolean doS1;
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
term = s1;
doS1 = true;
} else {
@@ -174,7 +174,7 @@ public class TestLongPostings extends Lu
int docID = -1;
while(docID < DocIdSetIterator.NO_MORE_DOCS) {
- final int what = random.nextInt(3);
+ final int what = random().nextInt(3);
if (what == 0) {
if (VERBOSE) {
System.out.println("TEST: docID=" + docID + "; do next()");
@@ -200,12 +200,12 @@ public class TestLongPostings extends Lu
break;
}
- if (random.nextInt(6) == 3) {
+ if (random().nextInt(6) == 3) {
final int freq = postings.freq();
assertTrue(freq >=1 && freq <= 4);
for(int pos=0;pos<freq;pos++) {
assertEquals(pos, postings.nextPosition());
- if (random.nextBoolean() && postings.hasPayload()) {
+ if (random().nextBoolean() && postings.hasPayload()) {
postings.getPayload();
}
}
@@ -214,9 +214,9 @@ public class TestLongPostings extends Lu
// advance
final int targetDocID;
if (docID == -1) {
- targetDocID = random.nextInt(NUM_DOCS+1);
+ targetDocID = random().nextInt(NUM_DOCS+1);
} else {
- targetDocID = docID + _TestUtil.nextInt(random, 1, NUM_DOCS - docID);
+ targetDocID = docID + _TestUtil.nextInt(random(), 1, NUM_DOCS - docID);
}
if (VERBOSE) {
System.out.println("TEST: docID=" + docID + "; do advance(" + targetDocID + ")");
@@ -242,12 +242,12 @@ public class TestLongPostings extends Lu
break;
}
- if (random.nextInt(6) == 3) {
+ if (random().nextInt(6) == 3) {
final int freq = postings.freq();
assertTrue(freq >=1 && freq <= 4);
for(int pos=0;pos<freq;pos++) {
assertEquals(pos, postings.nextPosition());
- if (random.nextBoolean() && postings.hasPayload()) {
+ if (random().nextBoolean() && postings.hasPayload()) {
postings.getPayload();
}
}
@@ -268,7 +268,7 @@ public class TestLongPostings extends Lu
public void doTestLongPostingsNoPositions(IndexOptions options) throws Exception {
// Don't use _TestUtil.getTempDir so that we own the
// randomness (ie same seed will point to same dir):
- Directory dir = newFSDirectory(_TestUtil.getTempDir("longpostings" + "." + random.nextLong()));
+ Directory dir = newFSDirectory(_TestUtil.getTempDir("longpostings" + "." + random().nextLong()));
final int NUM_DOCS = atLeast(2000);
@@ -293,19 +293,19 @@ public class TestLongPostings extends Lu
final FixedBitSet isS1 = new FixedBitSet(NUM_DOCS);
for(int idx=0;idx<NUM_DOCS;idx++) {
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
isS1.set(idx);
}
}
final IndexReader r;
if (true) {
- final IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ final IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setOpenMode(IndexWriterConfig.OpenMode.CREATE)
.setMergePolicy(newLogMergePolicy());
- iwc.setRAMBufferSizeMB(16.0 + 16.0 * random.nextDouble());
+ iwc.setRAMBufferSizeMB(16.0 + 16.0 * random().nextDouble());
iwc.setMaxBufferedDocs(-1);
- final RandomIndexWriter riw = new RandomIndexWriter(random, dir, iwc);
+ final RandomIndexWriter riw = new RandomIndexWriter(random(), dir, iwc);
FieldType ft = new FieldType(TextField.TYPE_UNSTORED);
ft.setIndexOptions(options);
@@ -313,7 +313,7 @@ public class TestLongPostings extends Lu
final Document doc = new Document();
String s = isS1.get(idx) ? s1 : s2;
final Field f = newField("field", s, ft);
- final int count = _TestUtil.nextInt(random, 1, 4);
+ final int count = _TestUtil.nextInt(random(), 1, 4);
for(int ct=0;ct<count;ct++) {
doc.add(f);
}
@@ -352,7 +352,7 @@ public class TestLongPostings extends Lu
final String term;
final boolean doS1;
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
term = s1;
doS1 = true;
} else {
@@ -368,17 +368,17 @@ public class TestLongPostings extends Lu
final DocsEnum postings;
if (options == IndexOptions.DOCS_ONLY) {
- docs = _TestUtil.docs(random, r, "field", new BytesRef(term), null, null, false);
+ docs = _TestUtil.docs(random(), r, "field", new BytesRef(term), null, null, false);
postings = null;
} else {
- docs = postings = _TestUtil.docs(random, r, "field", new BytesRef(term), null, null, true);
+ docs = postings = _TestUtil.docs(random(), r, "field", new BytesRef(term), null, null, true);
assert postings != null;
}
assert docs != null;
int docID = -1;
while(docID < DocIdSetIterator.NO_MORE_DOCS) {
- final int what = random.nextInt(3);
+ final int what = random().nextInt(3);
if (what == 0) {
if (VERBOSE) {
System.out.println("TEST: docID=" + docID + "; do next()");
@@ -404,7 +404,7 @@ public class TestLongPostings extends Lu
break;
}
- if (random.nextInt(6) == 3 && postings != null) {
+ if (random().nextInt(6) == 3 && postings != null) {
final int freq = postings.freq();
assertTrue(freq >=1 && freq <= 4);
}
@@ -412,9 +412,9 @@ public class TestLongPostings extends Lu
// advance
final int targetDocID;
if (docID == -1) {
- targetDocID = random.nextInt(NUM_DOCS+1);
+ targetDocID = random().nextInt(NUM_DOCS+1);
} else {
- targetDocID = docID + _TestUtil.nextInt(random, 1, NUM_DOCS - docID);
+ targetDocID = docID + _TestUtil.nextInt(random(), 1, NUM_DOCS - docID);
}
if (VERBOSE) {
System.out.println("TEST: docID=" + docID + "; do advance(" + targetDocID + ")");
@@ -440,7 +440,7 @@ public class TestLongPostings extends Lu
break;
}
- if (random.nextInt(6) == 3 && postings != null) {
+ if (random().nextInt(6) == 3 && postings != null) {
final int freq = postings.freq();
assertTrue("got invalid freq=" + freq, freq >=1 && freq <= 4);
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMaxTermFrequency.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMaxTermFrequency.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMaxTermFrequency.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMaxTermFrequency.java Sun Apr 15 14:41:44 2012
@@ -46,9 +46,9 @@ public class TestMaxTermFrequency extend
super.setUp();
dir = newDirectory();
IndexWriterConfig config = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random, MockTokenizer.SIMPLE, true)).setMergePolicy(newLogMergePolicy());
+ new MockAnalyzer(random(), MockTokenizer.SIMPLE, true)).setMergePolicy(newLogMergePolicy());
config.setSimilarity(new TestSimilarity());
- RandomIndexWriter writer = new RandomIndexWriter(random, dir, config);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir, config);
Document doc = new Document();
Field foo = newField("foo", "", TextField.TYPE_UNSTORED);
doc.add(foo);
@@ -81,16 +81,16 @@ public class TestMaxTermFrequency extend
*/
private String addValue() {
List<String> terms = new ArrayList<String>();
- int maxCeiling = _TestUtil.nextInt(random, 0, 255);
+ int maxCeiling = _TestUtil.nextInt(random(), 0, 255);
int max = 0;
for (char ch = 'a'; ch <= 'z'; ch++) {
- int num = _TestUtil.nextInt(random, 0, maxCeiling);
+ int num = _TestUtil.nextInt(random(), 0, maxCeiling);
for (int i = 0; i < num; i++)
terms.add(Character.toString(ch));
max = Math.max(max, num);
}
expected.add(max);
- Collections.shuffle(terms, random);
+ Collections.shuffle(terms, random());
return Arrays.toString(terms.toArray(new String[terms.size()]));
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMixedCodecs.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMixedCodecs.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMixedCodecs.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMixedCodecs.java Sun Apr 15 14:41:44 2012
@@ -49,8 +49,8 @@ public class TestMixedCodecs extends Luc
System.out.println("TEST: " + docUpto + " of " + NUM_DOCS);
}
if (docsLeftInThisSegment == 0) {
- final IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
- if (random.nextBoolean()) {
+ final IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
+ if (random().nextBoolean()) {
// Make sure we aggressively mix in SimpleText
// since it has different impls for all codec
// formats...
@@ -59,8 +59,8 @@ public class TestMixedCodecs extends Luc
if (w != null) {
w.close();
}
- w = new RandomIndexWriter(random, dir, iwc);
- docsLeftInThisSegment = _TestUtil.nextInt(random, 10, 100);
+ w = new RandomIndexWriter(random(), dir, iwc);
+ docsLeftInThisSegment = _TestUtil.nextInt(random(), 10, 100);
}
final Document doc = new Document();
doc.add(newField("id", String.valueOf(docUpto), StringField.TYPE_STORED));
@@ -76,11 +76,11 @@ public class TestMixedCodecs extends Luc
// Random delete half the docs:
final Set<Integer> deleted = new HashSet<Integer>();
while(deleted.size() < NUM_DOCS/2) {
- final Integer toDelete = random.nextInt(NUM_DOCS);
+ final Integer toDelete = random().nextInt(NUM_DOCS);
if (!deleted.contains(toDelete)) {
deleted.add(toDelete);
w.deleteDocuments(new Term("id", String.valueOf(toDelete)));
- if (random.nextInt(17) == 6) {
+ if (random().nextInt(17) == 6) {
final IndexReader r = w.getReader();
assertEquals(NUM_DOCS - deleted.size(), r.numDocs());
r.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java Sun Apr 15 14:41:44 2012
@@ -36,34 +36,34 @@ public class TestMultiFields extends Luc
Directory dir = newDirectory();
- IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(NoMergePolicy.COMPOUND_FILES));
+ IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(NoMergePolicy.COMPOUND_FILES));
_TestUtil.keepFullyDeletedSegments(w);
Map<BytesRef,List<Integer>> docs = new HashMap<BytesRef,List<Integer>>();
Set<Integer> deleted = new HashSet<Integer>();
List<BytesRef> terms = new ArrayList<BytesRef>();
- int numDocs = _TestUtil.nextInt(random, 1, 100 * RANDOM_MULTIPLIER);
+ int numDocs = _TestUtil.nextInt(random(), 1, 100 * RANDOM_MULTIPLIER);
Document doc = new Document();
Field f = newField("field", "", StringField.TYPE_UNSTORED);
doc.add(f);
Field id = newField("id", "", StringField.TYPE_UNSTORED);
doc.add(id);
- boolean onlyUniqueTerms = random.nextBoolean();
+ boolean onlyUniqueTerms = random().nextBoolean();
if (VERBOSE) {
System.out.println("TEST: onlyUniqueTerms=" + onlyUniqueTerms + " numDocs=" + numDocs);
}
Set<BytesRef> uniqueTerms = new HashSet<BytesRef>();
for(int i=0;i<numDocs;i++) {
- if (!onlyUniqueTerms && random.nextBoolean() && terms.size() > 0) {
+ if (!onlyUniqueTerms && random().nextBoolean() && terms.size() > 0) {
// re-use existing term
- BytesRef term = terms.get(random.nextInt(terms.size()));
+ BytesRef term = terms.get(random().nextInt(terms.size()));
docs.get(term).add(i);
f.setStringValue(term.utf8ToString());
} else {
- String s = _TestUtil.randomUnicodeString(random, 10);
+ String s = _TestUtil.randomUnicodeString(random(), 10);
BytesRef term = new BytesRef(s);
if (!docs.containsKey(term)) {
docs.put(term, new ArrayList<Integer>());
@@ -75,11 +75,11 @@ public class TestMultiFields extends Luc
}
id.setStringValue(""+i);
w.addDocument(doc);
- if (random.nextInt(4) == 1) {
+ if (random().nextInt(4) == 1) {
w.commit();
}
- if (i > 0 && random.nextInt(20) == 1) {
- int delID = random.nextInt(i);
+ if (i > 0 && random().nextInt(20) == 1) {
+ int delID = random().nextInt(i);
deleted.add(delID);
w.deleteDocuments(new Term("id", ""+delID));
if (VERBOSE) {
@@ -116,12 +116,12 @@ public class TestMultiFields extends Luc
}
for(int i=0;i<100;i++) {
- BytesRef term = terms.get(random.nextInt(terms.size()));
+ BytesRef term = terms.get(random().nextInt(terms.size()));
if (VERBOSE) {
System.out.println("TEST: seek term="+ UnicodeUtil.toHexString(term.utf8ToString()) + " " + term);
}
- DocsEnum docsEnum = _TestUtil.docs(random, reader, "field", term, liveDocs, null, false);
+ DocsEnum docsEnum = _TestUtil.docs(random(), reader, "field", term, liveDocs, null, false);
assertNotNull(docsEnum);
for(int docID : docs.get(term)) {
@@ -154,7 +154,7 @@ public class TestMultiFields extends Luc
public void testSeparateEnums() throws Exception {
Directory dir = newDirectory();
- IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document d = new Document();
d.add(newField("f", "j", StringField.TYPE_UNSTORED));
w.addDocument(d);
@@ -162,8 +162,8 @@ public class TestMultiFields extends Luc
w.addDocument(d);
IndexReader r = w.getReader();
w.close();
- DocsEnum d1 = _TestUtil.docs(random, r, "f", new BytesRef("j"), null, null, false);
- DocsEnum d2 = _TestUtil.docs(random, r, "f", new BytesRef("j"), null, null, false);
+ DocsEnum d1 = _TestUtil.docs(random(), r, "f", new BytesRef("j"), null, null, false);
+ DocsEnum d2 = _TestUtil.docs(random(), r, "f", new BytesRef("j"), null, null, false);
assertEquals(0, d1.nextDoc());
assertEquals(0, d2.nextDoc());
r.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiLevelSkipList.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiLevelSkipList.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiLevelSkipList.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestMultiLevelSkipList.java Sun Apr 15 14:41:44 2012
@@ -48,7 +48,7 @@ public class TestMultiLevelSkipList exte
class CountingRAMDirectory extends MockDirectoryWrapper {
public CountingRAMDirectory(Directory delegate) {
- super(random, delegate);
+ super(random(), delegate);
}
@Override
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTReaderWithThreads.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTReaderWithThreads.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTReaderWithThreads.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTReaderWithThreads.java Sun Apr 15 14:41:44 2012
@@ -32,7 +32,7 @@ public class TestNRTReaderWithThreads ex
Directory mainDir = newDirectory();
IndexWriter writer = new IndexWriter(
mainDir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).
setMaxBufferedDocs(10).
setMergePolicy(newLogMergePolicy(false,2))
);
@@ -76,7 +76,7 @@ public class TestNRTReaderWithThreads ex
int delCount = 0;
int addCount = 0;
int type;
- final Random r = new Random(random.nextLong());
+ final Random r = new Random(random().nextLong());
public RunThread(int type, IndexWriter writer) {
this.type = type;
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTThreads.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTThreads.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTThreads.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNRTThreads.java Sun Apr 15 14:41:44 2012
@@ -39,7 +39,7 @@ public class TestNRTThreads extends Thre
DirectoryReader r = IndexReader.open(writer, true);
while (System.currentTimeMillis() < stopTime && !failed.get()) {
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
if (VERBOSE) {
System.out.println("TEST: now reopen r=" + r);
}
@@ -106,7 +106,7 @@ public class TestNRTThreads extends Thre
@Override
protected IndexSearcher getFinalSearcher() throws Exception {
final IndexReader r2;
- if (random.nextBoolean()) {
+ if (random().nextBoolean()) {
r2 = writer.getReader();
} else {
writer.commit();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNeverDelete.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNeverDelete.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNeverDelete.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNeverDelete.java Sun Apr 15 14:41:44 2012
@@ -42,15 +42,15 @@ public class TestNeverDelete extends Luc
// them. This is still worth running on Windows since
// some files the IR opens and closes.
d.setNoDeleteOpenFile(false);
- final RandomIndexWriter w = new RandomIndexWriter(random,
+ final RandomIndexWriter w = new RandomIndexWriter(random(),
d,
newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random))
+ new MockAnalyzer(random()))
.setIndexDeletionPolicy(NoDeletionPolicy.INSTANCE));
- w.w.getConfig().setMaxBufferedDocs(_TestUtil.nextInt(random, 5, 30));
+ w.w.getConfig().setMaxBufferedDocs(_TestUtil.nextInt(random(), 5, 30));
w.commit();
- Thread[] indexThreads = new Thread[random.nextInt(4)];
+ Thread[] indexThreads = new Thread[random().nextInt(4)];
final long stopTime = System.currentTimeMillis() + atLeast(1000);
for (int x=0; x < indexThreads.length; x++) {
indexThreads[x] = new Thread() {
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNewestSegment.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNewestSegment.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNewestSegment.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNewestSegment.java Sun Apr 15 14:41:44 2012
@@ -24,7 +24,7 @@ import org.apache.lucene.util.LuceneTest
public class TestNewestSegment extends LuceneTestCase {
public void testNewestSegment() throws Exception {
Directory directory = newDirectory();
- IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random())));
assertNull(writer.newestSegment());
writer.close();
directory.close();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNoDeletionPolicy.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNoDeletionPolicy.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNoDeletionPolicy.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNoDeletionPolicy.java Sun Apr 15 14:41:44 2012
@@ -69,7 +69,7 @@ public class TestNoDeletionPolicy extend
public void testAllCommitsRemain() throws Exception {
Directory dir = newDirectory();
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(
- TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ TEST_VERSION_CURRENT, new MockAnalyzer(random()))
.setIndexDeletionPolicy(NoDeletionPolicy.INSTANCE));
for (int i = 0; i < 10; i++) {
Document doc = new Document();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNorms.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNorms.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNorms.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestNorms.java Sun Apr 15 14:41:44 2012
@@ -18,6 +18,7 @@ package org.apache.lucene.index;
*/
import java.io.IOException;
+import java.util.Random;
import org.apache.lucene.analysis.MockAnalyzer;
import org.apache.lucene.document.Document;
@@ -59,9 +60,9 @@ public class TestNorms extends LuceneTes
// LUCENE-1260
public void testCustomEncoder() throws Exception {
Directory dir = newDirectory();
- IndexWriterConfig config = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random));
+ IndexWriterConfig config = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random()));
config.setSimilarity(new CustomNormEncodingSimilarity());
- RandomIndexWriter writer = new RandomIndexWriter(random, dir, config);
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir, config);
Document doc = new Document();
Field foo = newField("foo", "", TextField.TYPE_UNSTORED);
Field bar = newField("bar", "", TextField.TYPE_UNSTORED);
@@ -115,11 +116,11 @@ public class TestNorms extends LuceneTes
*/
public void testNormsNotPresent() throws IOException {
Directory dir = newDirectory();
- boolean firstWriteNorm = random.nextBoolean();
+ boolean firstWriteNorm = random().nextBoolean();
buildIndex(dir, firstWriteNorm);
Directory otherDir = newDirectory();
- boolean secondWriteNorm = random.nextBoolean();
+ boolean secondWriteNorm = random().nextBoolean();
buildIndex(otherDir, secondWriteNorm);
AtomicReader reader = SlowCompositeReaderWrapper.wrap(IndexReader.open(otherDir));
@@ -134,8 +135,8 @@ public class TestNorms extends LuceneTes
}
IndexWriterConfig config = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random));
- RandomIndexWriter writer = new RandomIndexWriter(random, dir, config);
+ new MockAnalyzer(random()));
+ RandomIndexWriter writer = new RandomIndexWriter(random(), dir, config);
writer.addIndexes(reader);
AtomicReader mergedReader = SlowCompositeReaderWrapper.wrap(writer.getReader());
if (!firstWriteNorm && !secondWriteNorm) {
@@ -173,8 +174,9 @@ public class TestNorms extends LuceneTes
public void buildIndex(Directory dir, boolean writeNorms) throws IOException,
CorruptIndexException {
+ Random random = random();
IndexWriterConfig config = newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random));
+ new MockAnalyzer(random()));
Similarity provider = new MySimProvider(writeNorms);
config.setSimilarity(provider);
RandomIndexWriter writer = new RandomIndexWriter(random, dir, config);
@@ -182,7 +184,7 @@ public class TestNorms extends LuceneTes
int num = atLeast(100);
for (int i = 0; i < num; i++) {
Document doc = docs.nextDoc();
- int boost = writeNorms ? 1 + random.nextInt(255) : 0;
+ int boost = writeNorms ? 1 + random().nextInt(255) : 0;
Field f = new Field(byteTestField, "" + boost,
TextField.TYPE_STORED);
f.setBoost(boost);
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitNorms.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitNorms.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitNorms.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitNorms.java Sun Apr 15 14:41:44 2012
@@ -34,7 +34,7 @@ public class TestOmitNorms extends Lucen
// omitNorms bit in the FieldInfo
public void testOmitNorms() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer));
Document d = new Document();
@@ -79,7 +79,7 @@ public class TestOmitNorms extends Lucen
// omitNorms for the same field works
public void testMixedMerge() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(
ram,
newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer).
@@ -134,7 +134,7 @@ public class TestOmitNorms extends Lucen
// field,
public void testMixedRAM() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(
ram,
newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer).
@@ -188,7 +188,7 @@ public class TestOmitNorms extends Lucen
// Verifies no *.nrm exists when all fields omit norms:
public void testNoNrmFile() throws Throwable {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig(
TEST_VERSION_CURRENT, analyzer).setMaxBufferedDocs(3).setMergePolicy(newLogMergePolicy()));
LogMergePolicy lmp = (LogMergePolicy) writer.getConfig().getMergePolicy();
@@ -265,10 +265,10 @@ public class TestOmitNorms extends Lucen
* Indexes at least 1 document with f1, and at least 1 document with f2.
* returns the norms for "field".
*/
- static byte[] getNorms(String field, Field f1, Field f2) throws IOException {
+ byte[] getNorms(String field, Field f1, Field f2) throws IOException {
Directory dir = newDirectory();
- IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy());
- RandomIndexWriter riw = new RandomIndexWriter(random, dir, iwc);
+ IndexWriterConfig iwc = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())).setMergePolicy(newLogMergePolicy());
+ RandomIndexWriter riw = new RandomIndexWriter(random(), dir, iwc);
// add f1
Document d = new Document();
@@ -281,10 +281,10 @@ public class TestOmitNorms extends Lucen
riw.addDocument(d);
// add a mix of f1's and f2's
- int numExtraDocs = _TestUtil.nextInt(random, 1, 1000);
+ int numExtraDocs = _TestUtil.nextInt(random(), 1, 1000);
for (int i = 0; i < numExtraDocs; i++) {
d = new Document();
- d.add(random.nextBoolean() ? f1 : f2);
+ d.add(random().nextBoolean() ? f1 : f2);
riw.addDocument(d);
}
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitPositions.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitPositions.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitPositions.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitPositions.java Sun Apr 15 14:41:44 2012
@@ -38,7 +38,7 @@ public class TestOmitPositions extends L
public void testBasic() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter w = new RandomIndexWriter(random, dir);
+ RandomIndexWriter w = new RandomIndexWriter(random(), dir);
Document doc = new Document();
FieldType ft = new FieldType(TextField.TYPE_UNSTORED);
ft.setIndexOptions(IndexOptions.DOCS_AND_FREQS);
@@ -53,7 +53,7 @@ public class TestOmitPositions extends L
assertNull(MultiFields.getTermPositionsEnum(reader, null, "foo", new BytesRef("test"), false));
- DocsEnum de = _TestUtil.docs(random, reader, "foo", new BytesRef("test"), null, null, true);
+ DocsEnum de = _TestUtil.docs(random(), reader, "foo", new BytesRef("test"), null, null, true);
while (de.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) {
assertEquals(2, de.freq());
}
@@ -66,7 +66,7 @@ public class TestOmitPositions extends L
// omitTermFreqAndPositions bit in the FieldInfo
public void testPositions() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer));
Document d = new Document();
@@ -189,7 +189,7 @@ public class TestOmitPositions extends L
// Verifies no *.prx exists when all fields omit term positions:
public void testNoPrxFile() throws Throwable {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig(
TEST_VERSION_CURRENT, analyzer).setMaxBufferedDocs(3).setMergePolicy(newLogMergePolicy()));
LogMergePolicy lmp = (LogMergePolicy) writer.getConfig().getMergePolicy();
Modified: lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitTf.java
URL: http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitTf.java?rev=1326351&r1=1326350&r2=1326351&view=diff
==============================================================================
--- lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitTf.java (original)
+++ lucene/dev/trunk/lucene/core/src/test/org/apache/lucene/index/TestOmitTf.java Sun Apr 15 14:41:44 2012
@@ -61,7 +61,7 @@ public class TestOmitTf extends LuceneTe
// omitTermFreqAndPositions bit in the FieldInfo
public void testOmitTermFreqAndPositions() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer));
Document d = new Document();
@@ -106,7 +106,7 @@ public class TestOmitTf extends LuceneTe
// omitTermFreqAndPositions for the same field works
public void testMixedMerge() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(
ram,
newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer).
@@ -159,7 +159,7 @@ public class TestOmitTf extends LuceneTe
// field,
public void testMixedRAM() throws Exception {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(
ram,
newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer).
@@ -208,7 +208,7 @@ public class TestOmitTf extends LuceneTe
// Verifies no *.prx exists when all fields omit term freq:
public void testNoPrxFile() throws Throwable {
Directory ram = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig(
TEST_VERSION_CURRENT, analyzer).setMaxBufferedDocs(3).setMergePolicy(newLogMergePolicy()));
LogMergePolicy lmp = (LogMergePolicy) writer.getConfig().getMergePolicy();
@@ -247,7 +247,7 @@ public class TestOmitTf extends LuceneTe
// Test scores with one field with Term Freqs and one without, otherwise with equal content
public void testBasic() throws Exception {
Directory dir = newDirectory();
- Analyzer analyzer = new MockAnalyzer(random);
+ Analyzer analyzer = new MockAnalyzer(random());
IndexWriter writer = new IndexWriter(
dir,
newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer).
@@ -427,8 +427,8 @@ public class TestOmitTf extends LuceneTe
/** test that when freqs are omitted, that totalTermFreq and sumTotalTermFreq are -1 */
public void testStats() throws Exception {
Directory dir = newDirectory();
- RandomIndexWriter iw = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)));
+ RandomIndexWriter iw = new RandomIndexWriter(random(), dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random())));
Document doc = new Document();
FieldType ft = new FieldType(TextField.TYPE_UNSTORED);
ft.setIndexOptions(IndexOptions.DOCS_ONLY);