You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by ec...@apache.org on 2013/07/12 17:11:14 UTC
svn commit: r1502583 - in /accumulo/branches/1.5/test:
src/main/java/org/apache/accumulo/test/CreateRFiles.java
src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java
system/auto/simple/bulk.py system/auto/simple/compaction.py
Author: ecn
Date: Fri Jul 12 15:11:14 2013
New Revision: 1502583
URL: http://svn.apache.org/r1502583
Log:
ACCUMULO-1563 committing Jonathan Hsieh's patch to prevent the test from writing to /
Modified:
accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java
accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java
accumulo/branches/1.5/test/system/auto/simple/bulk.py
accumulo/branches/1.5/test/system/auto/simple/compaction.py
Modified: accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java
URL: http://svn.apache.org/viewvc/accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java?rev=1502583&r1=1502582&r2=1502583&view=diff
==============================================================================
--- accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java (original)
+++ accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java Fri Jul 12 15:11:14 2013
@@ -58,7 +58,7 @@ public class CreateRFiles {
int count = 0;
while (currEnd <= opts.end && currStart < currEnd) {
- final String tia = String.format("--rfile /%s/mf%05d --timestamp 1 --size 50 --random 56 --rows %d --start %d --user root", opts.outputDirectory, count, currEnd - currStart, currStart);
+ final String tia = String.format("--rfile %s/mf%05d --timestamp 1 --size 50 --random 56 --rows %d --start %d --user root", opts.outputDirectory, count, currEnd - currStart, currStart);
Runnable r = new Runnable() {
Modified: accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java
URL: http://svn.apache.org/viewvc/accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java?rev=1502583&r1=1502582&r2=1502583&view=diff
==============================================================================
--- accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java (original)
+++ accumulo/branches/1.5/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java Fri Jul 12 15:11:14 2013
@@ -44,8 +44,8 @@ public class BulkSplitOptimizationTest e
@Override
public void cleanup() throws Exception {
FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
- fs.delete(new Path("/tmp/testmf"), true);
- fs.delete(new Path("/tmp/testmf_failures"), true);
+ fs.delete(new Path("tmp/testmf"), true);
+ fs.delete(new Path("tmp/testmf_failures"), true);
}
@Override
@@ -63,11 +63,11 @@ public class BulkSplitOptimizationTest e
public void run() throws Exception {
FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
- fs.delete(new Path("/tmp/testmf"), true);
+ fs.delete(new Path("tmp/testmf"), true);
AuthenticationToken token = this.getToken();
CreateRFiles.main(new String[] {"--output", "tmp/testmf", "--numThreads", "8", "--start", "0", "--end", "100000", "--splits", "99"});
- bulkImport(fs, TABLE_NAME, "/tmp/testmf");
+ bulkImport(fs, TABLE_NAME, "tmp/testmf");
checkSplits(TABLE_NAME, 0, 0);
checkRFiles(TABLE_NAME, 1, 1, 100, 100);
Modified: accumulo/branches/1.5/test/system/auto/simple/bulk.py
URL: http://svn.apache.org/viewvc/accumulo/branches/1.5/test/system/auto/simple/bulk.py?rev=1502583&r1=1502582&r2=1502583&view=diff
==============================================================================
--- accumulo/branches/1.5/test/system/auto/simple/bulk.py (original)
+++ accumulo/branches/1.5/test/system/auto/simple/bulk.py Fri Jul 12 15:11:14 2013
@@ -42,13 +42,13 @@ class SimpleBulkTest(TestUtilsMixin, uni
self.masterHost(),
'org.apache.accumulo.test.BulkImportDirectory',
['-u', ROOT, '-p', ROOT_PASSWORD,
- '-t', 'test_ingest', '--source', dir, '--failures', '/testBulkFail', '-i', INSTANCE_NAME])
+ '-t', 'test_ingest', '--source', dir, '--failures', 'testBulkFail', '-i', INSTANCE_NAME])
self.wait(handle)
self.assert_(handle.returncode == 0)
def createRFiles(self):
- args = '--rfile /testrf/rf%02d --timestamp 1 --size 50 --random 56 --rows %1d --start %ld --cols 1 -u root -i ' + INSTANCE_NAME
+ args = '--rfile testrf/rf%02d --timestamp 1 --size 50 --random 56 --rows %1d --start %ld --cols 1 -u root -i ' + INSTANCE_NAME
log.info('creating rfiles')
handles = []
for i in range(COUNT):
@@ -73,9 +73,9 @@ class SimpleBulkTest(TestUtilsMixin, uni
# initialize the database
self.createTable('test_ingest')
- self.execute(self.masterHost(), 'hadoop dfs -rmr /testrf'.split())
- self.execute(self.masterHost(), 'hadoop dfs -rmr /testBulkFail'.split())
- self.execute(self.masterHost(), 'hadoop dfs -mkdir /testBulkFail'.split())
+ self.execute(self.masterHost(), 'hadoop dfs -rmr testrf'.split())
+ self.execute(self.masterHost(), 'hadoop dfs -rmr testBulkFail'.split())
+ self.execute(self.masterHost(), 'hadoop dfs -mkdir testBulkFail'.split())
# insert some data
self.createRFiles()
Modified: accumulo/branches/1.5/test/system/auto/simple/compaction.py
URL: http://svn.apache.org/viewvc/accumulo/branches/1.5/test/system/auto/simple/compaction.py?rev=1502583&r1=1502582&r2=1502583&view=diff
==============================================================================
--- accumulo/branches/1.5/test/system/auto/simple/compaction.py (original)
+++ accumulo/branches/1.5/test/system/auto/simple/compaction.py Fri Jul 12 15:11:14 2013
@@ -52,8 +52,8 @@ class CompactionTest(SimpleBulkTest):
# initialize the database
self.createTable('test_ingest')
- self.execute(self.masterHost(), 'hadoop dfs -rmr /testrf'.split())
- self.execute(self.masterHost(), 'hadoop dfs -rmr /testrfFail'.split())
+ self.execute(self.masterHost(), 'hadoop dfs -rmr testrf'.split())
+ self.execute(self.masterHost(), 'hadoop dfs -rmr testrfFail'.split())
# insert some data
self.createRFiles(self.masterHost())