You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by to...@apache.org on 2011/04/07 01:58:51 UTC

svn commit: r1089684 - in /hbase/trunk: CHANGES.txt src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java

Author: todd
Date: Wed Apr  6 23:58:50 2011
New Revision: 1089684

URL: http://svn.apache.org/viewvc?rev=1089684&view=rev
Log:
HBASE-3746  Clean up CompressionTest to not directly reference DistributedFileSystem

Modified:
    hbase/trunk/CHANGES.txt
    hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java

Modified: hbase/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hbase/trunk/CHANGES.txt?rev=1089684&r1=1089683&r2=1089684&view=diff
==============================================================================
--- hbase/trunk/CHANGES.txt (original)
+++ hbase/trunk/CHANGES.txt Wed Apr  6 23:58:50 2011
@@ -169,6 +169,12 @@ Release 0.91.0 - Unreleased
    HBASE-3488  Add CellCounter to count multiple versions of rows
                (Subbu M. Iyer via Stack)
 
+Release 0.90.3 - Unreleased
+
+  BUG FIXES
+   HBASE-3746  Clean up CompressionTest to not directly reference
+               DistributedFileSystem (todd)
+
 
 Release 0.90.2 - Unreleased
   

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java?rev=1089684&r1=1089683&r2=1089684&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java Wed Apr  6 23:58:50 2011
@@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.util;
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hbase.io.hfile.Compression;
 import org.apache.hadoop.hbase.io.hfile.HFile;
@@ -92,55 +93,48 @@ public class CompressionTest {
   protected static Path path = new Path(".hfile-comp-test");
 
   public static void usage() {
-    System.err.println("Usage: CompressionTest HDFS_PATH none|gz|lzo");
+    System.err.println(
+      "Usage: CompressionTest <path> none|gz|lzo\n" +
+      "\n" +
+      "For example:\n" +
+      "  hbase " + CompressionTest.class + " file:///tmp/testfile gz\n");
     System.exit(1);
   }
 
-  protected static DistributedFileSystem openConnection(String urlString)
-  throws java.net.URISyntaxException, java.io.IOException {
-    URI dfsUri = new URI(urlString);
-    Configuration dfsConf = new Configuration();
-    DistributedFileSystem dfs = new DistributedFileSystem();
-    dfs.initialize(dfsUri, dfsConf);
-    return dfs;
-  }
-
-  protected static boolean closeConnection(DistributedFileSystem dfs) {
-    if (dfs != null) {
-      try {
-        dfs.close();
-      } catch (Exception e) {
-        e.printStackTrace();
-      }
+  public static void doSmokeTest(FileSystem fs, Path path, String codec)
+  throws Exception {
+    HFile.Writer writer = new HFile.Writer(
+      fs, path, HFile.DEFAULT_BLOCKSIZE, codec, null);
+    writer.append(Bytes.toBytes("testkey"), Bytes.toBytes("testval"));
+    writer.appendFileInfo(Bytes.toBytes("infokey"), Bytes.toBytes("infoval"));
+    writer.close();
+
+    HFile.Reader reader = new HFile.Reader(fs, path, null, false, false);
+    reader.loadFileInfo();
+    byte[] key = reader.getFirstKey();
+    boolean rc = Bytes.toString(key).equals("testkey");
+    reader.close();
+
+    if (!rc) {
+      throw new Exception("Read back incorrect result: " +
+                          Bytes.toStringBinary(key));
     }
-    return dfs == null;
   }
 
-  public static void main(String[] args) {
-    if (args.length != 2) usage();
+  public static void main(String[] args) throws Exception {
+    if (args.length != 2) {
+      usage();
+      System.exit(1);
+    }
+
+    Configuration conf = new Configuration();
+    Path path = new Path(args[0]);
+    FileSystem fs = path.getFileSystem(conf);
     try {
-      DistributedFileSystem dfs = openConnection(args[0]);
-      dfs.delete(path, false);
-      HFile.Writer writer = new HFile.Writer(dfs, path,
-        HFile.DEFAULT_BLOCKSIZE, args[1], null);
-      writer.append(Bytes.toBytes("testkey"), Bytes.toBytes("testval"));
-      writer.appendFileInfo(Bytes.toBytes("infokey"), Bytes.toBytes("infoval"));
-      writer.close();
-
-      HFile.Reader reader = new HFile.Reader(dfs, path, null, false, false);
-      reader.loadFileInfo();
-      byte[] key = reader.getFirstKey();
-      boolean rc = Bytes.toString(key).equals("testkey");
-      reader.close();
-
-      dfs.delete(path, false);
-      closeConnection(dfs);
-
-      if (rc) System.exit(0);
-    } catch (Exception e) {
-      e.printStackTrace();
+      doSmokeTest(fs, path, args[1]);
+    } finally {
+      fs.delete(path, false);
     }
-    System.out.println("FAILED");
-    System.exit(1);
+    System.out.println("SUCCESS");
   }
 }