You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cu...@apache.org on 2007/04/19 23:34:53 UTC
svn commit: r530556 [12/12] - in /lucene/hadoop/trunk: ./
src/contrib/abacus/src/java/org/apache/hadoop/abacus/
src/contrib/hbase/src/java/org/apache/hadoop/hbase/
src/contrib/hbase/src/test/org/apache/hadoop/hbase/
src/contrib/streaming/src/java/org/a...
Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java Thu Apr 19 14:34:41 2007
@@ -196,8 +196,8 @@
public InputSplit[] getSplits(JobConf job,
int numSplits) throws IOException {
- return new MySplit[]{new MySplit(0,1), new MySplit(1,3),
- new MySplit(4,2)};
+ return new MySplit[]{new MySplit(0, 1), new MySplit(1, 3),
+ new MySplit(4, 2)};
}
public RecordReader getRecordReader(InputSplit split,
Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java Thu Apr 19 14:34:41 2007
@@ -73,7 +73,7 @@
InputFormat format = new SequenceFileInputFilter();
Text key = new Text();
BytesWritable value = new BytesWritable();
- if(numSplits==0) {
+ if (numSplits==0) {
numSplits =
random.nextInt(MAX_LENGTH/(SequenceFile.SYNC_INTERVAL/20))+1;
}
@@ -138,7 +138,7 @@
int count = countRecords(1);
LOG.info("Accepted "+count+" records");
int expectedCount = length/1000;
- if(expectedCount*1000!=length)
+ if (expectedCount*1000!=length)
expectedCount++;
assertEquals(count, expectedCount);
}
Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java Thu Apr 19 14:34:41 2007
@@ -21,15 +21,15 @@
static {
for(int i=0; i<dataNodes.length; i++) {
- cluster.add( dataNodes[i] );
+ cluster.add(dataNodes[i]);
}
}
public void testContains() {
for(int i=0; i<dataNodes.length; i++) {
- assertTrue( cluster.contains(dataNodes[i]));
+ assertTrue(cluster.contains(dataNodes[i]));
}
- assertFalse( cluster.contains( NODE ));
+ assertFalse(cluster.contains(NODE));
}
public void testNumOfChildren() throws Exception {
@@ -55,14 +55,14 @@
public void testRemove() throws Exception {
for(int i=0; i<dataNodes.length; i++) {
- cluster.remove( dataNodes[i] );
+ cluster.remove(dataNodes[i]);
}
for(int i=0; i<dataNodes.length; i++) {
- assertFalse( cluster.contains( dataNodes[i] ) );
+ assertFalse(cluster.contains(dataNodes[i]));
}
assertEquals(0, cluster.getNumOfLeaves());
for(int i=0; i<dataNodes.length; i++) {
- cluster.add( dataNodes[i] );
+ cluster.add(dataNodes[i]);
}
}
}
Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java Thu Apr 19 14:34:41 2007
@@ -287,7 +287,7 @@
fs.delete(randomOuts);
- JobConf genJob = new JobConf(conf,TestRecordMR.class);
+ JobConf genJob = new JobConf(conf, TestRecordMR.class);
genJob.setInputPath(randomIns);
genJob.setInputKeyClass(RecInt.class);
genJob.setInputValueClass(RecInt.class);
@@ -334,7 +334,7 @@
int intermediateReduces = 10;
Path intermediateOuts = new Path(testdir, "intermediateouts");
fs.delete(intermediateOuts);
- JobConf checkJob = new JobConf(conf,TestRecordMR.class);
+ JobConf checkJob = new JobConf(conf, TestRecordMR.class);
checkJob.setInputPath(randomOuts);
checkJob.setInputKeyClass(RecInt.class);
checkJob.setInputValueClass(RecString.class);
@@ -359,7 +359,7 @@
//
Path finalOuts = new Path(testdir, "finalouts");
fs.delete(finalOuts);
- JobConf mergeJob = new JobConf(conf,TestRecordMR.class);
+ JobConf mergeJob = new JobConf(conf, TestRecordMR.class);
mergeJob.setInputPath(intermediateOuts);
mergeJob.setInputKeyClass(RecInt.class);
mergeJob.setInputValueClass(RecString.class);
@@ -392,12 +392,12 @@
if (dist[i] == 0) {
continue;
}
- if (! in.next(key, val)) {
+ if (!in.next(key, val)) {
System.err.println("Cannot read entry " + i);
success = false;
break;
} else {
- if ( !((key.getData() == i ) && (val.getData() == dist[i]))) {
+ if (!((key.getData() == i) && (val.getData() == dist[i]))) {
System.err.println("Mismatch! Pos=" + key.getData() + ", i=" + i + ", val=" + val.getData() + ", dist[i]=" + dist[i]);
success = false;
}