You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pig.apache.org by ra...@apache.org on 2009/10/09 07:04:11 UTC

svn commit: r823411 [2/2] - in /hadoop/pig/trunk/contrib/zebra: ./ src/java/org/apache/hadoop/zebra/io/ src/java/org/apache/hadoop/zebra/types/ src/test/org/apache/hadoop/zebra/io/ src/test/org/apache/hadoop/zebra/pig/ src/test/org/apache/hadoop/zebra/...

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java Fri Oct  9 05:04:10 2009
@@ -69,8 +69,8 @@
   public void testInsertNullValues() throws IOException, ParseException {
     fs.delete(path, true);
     System.out.println("testInsertNullValues");
-    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part1", true);
     // Tuple row = TypesUtils.createTuple(writer.getSchema());
     // ins.insert(new BytesWritable("key".getBytes()), row);
@@ -82,8 +82,8 @@
   public void testFailureInvalidSchema() throws IOException, ParseException {
     fs.delete(path, true);
     System.out.println("testFailureInvalidSchema");
-    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part1", true);
     Tuple row = TypesUtils.createTuple(Schema.parse("xyz, ijk, def"));
     try {
@@ -102,8 +102,8 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testFailureGetInserterAfterWriterClosed");
-    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     try {
       writer.close();
       TableInserter ins = writer.getInserter("part1", true);
@@ -120,8 +120,8 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testFailureInsertAfterClose");
-    writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part1", true);
 
     Tuple row = TypesUtils.createTuple(writer.getSchema());
@@ -151,7 +151,7 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testFailureInsertXtraColumn");
-    writer = new ColumnGroup.Writer(path, "abc ", false, "pig", "lzo2", true,
+    writer = new ColumnGroup.Writer(path, "abc ", false, "pig", "gz", null, null, (short) -1, true,
         conf);
     TableInserter ins = writer.getInserter("part1", true);
 
@@ -164,7 +164,7 @@
       row.set(1, map);
       Assert
           .fail("Failed to catch insertion an extra column not defined in schema.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // noop, expecting exceptions
     } finally {
       ins.close();
@@ -177,8 +177,8 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testInsertOneRow");
-    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part1", true);
 
     Tuple row = TypesUtils.createTuple(writer.getSchema());
@@ -199,8 +199,8 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testInsert2Rows");
-    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part1", true);
 
     // row 1
@@ -233,8 +233,8 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testInsert2Inserters");
-    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins1 = writer.getInserter("part1", true);
     TableInserter ins2 = writer.getInserter("part2", true);
 
@@ -270,8 +270,8 @@
       ParseException {
     fs.delete(path, true);
     System.out.println("testFailureOverlappingKeys");
-    writer = new ColumnGroup.Writer(path, "abc, def ", true, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def ", true, "pig", "gz",
+        null, null, (short) -1, true, conf);
     TableInserter ins1 = writer.getInserter("part1", false);
     TableInserter ins2 = writer.getInserter("part2", false);
 

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java Fri Oct  9 05:04:10 2009
@@ -62,8 +62,8 @@
   @Test
   public void testNew() throws IOException, ParseException {
     System.out.println("testNew");
-    writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2",
-        true, conf);
+    writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz",
+        null, null, (short) -1, true, conf);
     // NOTE: don't call writer.close() here
     close();
   }
@@ -73,11 +73,11 @@
       ParseException {
     System.out.println("testFailureExistingSortedDiff");
     try {
-      writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2",
-          true, conf);
+      writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz",
+          null, null, (short) -1, true, conf);
       finish();
-      writer = new ColumnGroup.Writer(path, "abc, def", true, "pig", "lzo2",
-          false, conf);
+      writer = new ColumnGroup.Writer(path, "abc, def", true, "pig", "gz",
+          null, null, (short) -1, false, conf);
       Assert.fail("Failed to catch sorted flag alteration.");
     } catch (IOException e) {
       // noop, expecting exceptions
@@ -89,8 +89,8 @@
   @Test
   public void testExisting() throws IOException, ParseException {
     System.out.println("testExisting");
-    writer = new ColumnGroup.Writer(path, " abc ,  def ", false, "pig", "lzo2",
-        false, conf);
+    writer = new ColumnGroup.Writer(path, " abc ,  def ", false, "pig", "gz",
+        null, null, (short) -1, false, conf);
     writer.close();
     close();
   }
@@ -105,7 +105,7 @@
       FSDataOutputStream in = fs.create(path);
       in.close();
       writer = new ColumnGroup.Writer(path, "   abc ,  def   ", false, "pig",
-          "lzo2", false, conf);
+          "gz", null, null, (short) -1, false, conf);
       Assert.fail("Failed to catch path not a directory.");
     } catch (IOException e) {
       // noop, expecting exceptions
@@ -121,7 +121,7 @@
       fs.delete(path, true);
       FSDataOutputStream in = fs.create(new Path(path, ColumnGroup.META_FILE));
       in.close();
-      writer = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", false,
+      writer = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, false,
           conf);
       Assert.fail("Failed to catch meta file existence.");
     } catch (IOException e) {
@@ -135,10 +135,10 @@
   public void testFailureDiffSchema() throws IOException, ParseException {
     System.out.println("testFailureDiffSchema");
     try {
-      writer = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", false,
+      writer = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, false,
           conf);
       writer.finish();
-      writer = new ColumnGroup.Writer(path, "efg", false, "pig", "lzo2", false,
+      writer = new ColumnGroup.Writer(path, "efg", false, "pig", "gz", null, null, (short) -1, false,
           conf);
       Assert.fail("Failed to catch schema differences.");
     } catch (IOException e) {
@@ -155,7 +155,7 @@
     ColumnGroup.Writer writer2 = null;
     ColumnGroup.Writer writer3 = null;
     try {
-      writer1 = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", true,
+      writer1 = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, true,
           conf);
       writer2 = new ColumnGroup.Writer(path, conf);
       writer3 = new ColumnGroup.Writer(path, conf);

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java Fri Oct  9 05:04:10 2009
@@ -63,7 +63,7 @@
     schema = new Schema("a,b,c,d,e,f,g");
 
     ColumnGroup.Writer writer = new ColumnGroup.Writer(path, schema, false,
-        "pig", "lzo2", true, conf);
+        "pig", "gz", null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part0", true);
 
     // row 1
@@ -168,7 +168,7 @@
     try {
       row.get(0);
       Assert.fail("Failed to catch out of boundary exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -202,7 +202,7 @@
     try {
       row.get(1);
       Assert.fail("Failed to catch out of boundary exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -236,7 +236,7 @@
     try {
       row.get(1);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -270,7 +270,7 @@
     try {
       row.get(2);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -304,7 +304,7 @@
     try {
       row.get(2);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java Fri Oct  9 05:04:10 2009
@@ -77,7 +77,7 @@
     System.out.println("testInsert2Inserters");
     boolean sorted = false; // true;
     writer = new ColumnGroup.Writer(path, "col1, colTWO", sorted, "pig",
-        "lzo2", true, conf);
+        "gz", null, null, (short) -1, true, conf);
     TableInserter ins1 = writer.getInserter("part1", false);
     TableInserter ins2 = writer.getInserter("part2", false);
 
@@ -197,8 +197,8 @@
       throws IOException, ExecException, ParseException {
     ColumnGroup.Writer writer = null;
     if (schemaString != null) {
-      writer = new ColumnGroup.Writer(path, schemaString, false, "pig", "lzo2",
-          true, conf);
+      writer = new ColumnGroup.Writer(path, schemaString, false, "pig", "gz",
+          null, null, (short) -1, true, conf);
     } else {
       writer = new ColumnGroup.Writer(path, conf);
     }

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java Fri Oct  9 05:04:10 2009
@@ -73,7 +73,7 @@
     schema = new Schema(STR_SCHEMA);
 
     ColumnGroup.Writer writer = new ColumnGroup.Writer(path, schema, false,
-        "pig", "lzo2", true, conf);
+        "pig", "gz", null, null, (short) -1, true, conf);
     TableInserter ins = writer.getInserter("part0", true);
 
     Tuple row = TypesUtils.createTuple(schema);
@@ -273,7 +273,7 @@
     try {
       row.get(0);
       Assert.fail("Failed to catch out of boundary exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -306,7 +306,7 @@
     try {
       row.get(1);
       Assert.fail("Failed to catch out of boundary exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -340,7 +340,7 @@
     try {
       row.get(1);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -374,7 +374,7 @@
     try {
       row.get(2);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -415,7 +415,7 @@
     try {
       Object obj = tupColl.get(1);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
     tupColl = itorColl.next();
@@ -424,7 +424,7 @@
     try {
       row.get(3);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -474,7 +474,7 @@
     try {
       Object obj = tupColl.get(1);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
     tupColl = itorColl.next();
@@ -482,7 +482,7 @@
     try {
       Object obj = tupColl.get(1);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
     Assert.assertFalse(itorColl.hasNext());
@@ -492,7 +492,7 @@
     try {
       Object obj = row.get(4);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 
@@ -545,7 +545,7 @@
     try {
       row.get(3);
       Assert.fail("Failed to catch 'out of boundary' exceptions.");
-    } catch (ExecException e) {
+    } catch (IndexOutOfBoundsException e) {
       // no op, expecting out of bounds exceptions
     }
 

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java Fri Oct  9 05:04:10 2009
@@ -210,7 +210,7 @@
   @Test
   public void testRead1() throws IOException, ParseException {
     String query = "records = LOAD '" + path.toString()
-        + "' USING org.apache.pig.table.pig.TableLoader('c');";
+        + "' USING org.apache.hadoop.zebra.pig.TableLoader('c');";
     System.out.println(query);
     pigServer.registerQuery(query);
     Iterator<Tuple> it = pigServer.openIterator("records");

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java Fri Oct  9 05:04:10 2009
@@ -262,7 +262,7 @@
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Encountered \" \"[\" \"[ \"\" at line 1, column 1.";
+      String str = "Encountered \" \"{\" \"{ \"\" at line 1, column 4.";
       System.out.println(errMsg);
       System.out.println(str);
       Assert.assertEquals(errMsg.startsWith(str), true);

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java Fri Oct  9 05:04:10 2009
@@ -81,8 +81,15 @@
   @Test
   public void testStorageValid1() {
     try {
-      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip ; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo2 SERIALIZE BY avro";
-      Partition p = new Partition(schema.toString(), strStorage);
+        String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+//      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+//      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav group:data perm:0766; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+//      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav group:data perm:966; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+//      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+//      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:ggg SECURE BY group:fff; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+//      String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gauravj user:gauravj; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+
+    	Partition p = new Partition(schema.toString(), strStorage);
       CGSchema[] cgschemas = p.getCGSchemas();
 
       // 3 column group;
@@ -157,4 +164,4 @@
       Assert.assertTrue(false);
     }
   }
-}
\ No newline at end of file
+}

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java Fri Oct  9 05:04:10 2009
@@ -207,9 +207,9 @@
 
       Assert.assertEquals(cgs1.getCompressor(), "gzip");
       Assert.assertEquals(cgs1.getSerializer(), "pig");
-      Assert.assertEquals(cgs2.getCompressor(), "lzo2");
+      Assert.assertEquals(cgs2.getCompressor(), "lzo");
       Assert.assertEquals(cgs2.getSerializer(), "avro");
-      Assert.assertEquals(cgs3.getCompressor(), "lzo2");
+      Assert.assertEquals(cgs3.getCompressor(), "lzo");
       Assert.assertEquals(cgs3.getSerializer(), "pig");
 
       System.out.println("*********** Column Map **********");
@@ -249,4 +249,4 @@
     } catch (Exception e) {
     }
   }
-}
\ No newline at end of file
+}

Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java Fri Oct  9 05:04:10 2009
@@ -107,11 +107,11 @@
       Assert.assertEquals("f6", f32.name);
       Assert.assertEquals(ColumnType.BYTES, f32.type);
 
-      Assert.assertEquals(cgs1.getCompressor(), "lzo2");
+      Assert.assertEquals(cgs1.getCompressor(), "lzo");
       Assert.assertEquals(cgs1.getSerializer(), "pig");
       Assert.assertEquals(cgs2.getCompressor(), "gzip");
       Assert.assertEquals(cgs2.getSerializer(), "avro");
-      Assert.assertEquals(cgs3.getCompressor(), "lzo2");
+      Assert.assertEquals(cgs3.getCompressor(), "lzo");
       Assert.assertEquals(cgs3.getSerializer(), "pig");
 
       System.out.println("*********** Column Map **********");
@@ -204,10 +204,10 @@
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n    <EOF> \n    \"compress by\" ...\n    \"serialize by\" ...\n    \"[\" ...\n    ";
+      String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\n";
       System.out.println(errMsg);
       System.out.println(str);
-      Assert.assertEquals(errMsg, str);
+      Assert.assertEquals(errMsg.startsWith(str), true);
     }
   }
 
@@ -238,10 +238,10 @@
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n    <EOF> \n    \"compress by\" ...\n    \"serialize by\" ...\n    \"[\" ...\n    ";
+      String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n";
       System.out.println(errMsg);
       System.out.println(str);
-      Assert.assertEquals(errMsg, str);
+      Assert.assertEquals(errMsg.startsWith(str), true);
     }
   }
 
@@ -255,10 +255,10 @@
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Encountered \" \"[\" \"[ \"\" at line 1, column 1.\nWas expecting one of:\n    \"compress by\" ...\n    \"serialize by\" ...\n    ";
+      String str = "Encountered \"<EOF>\" at line 1, column 1.\nWas expecting one of:\n";
       System.out.println(errMsg);
       System.out.println(str);
-      Assert.assertEquals(errMsg, str);
+      Assert.assertEquals(errMsg.startsWith(str), true);
     }
   }
 
@@ -289,24 +289,24 @@
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Lexical error at line 1, column 2.  Encountered: <EOF> after : \"\"";
+      String str = "Encountered \" \":\" \": \"\" at line 1, column 1.\n";
       System.out.println(errMsg);
       System.out.println(str);
-      Assert.assertEquals(errMsg, str);
+      Assert.assertEquals(errMsg.startsWith(str), true);
     }
   }
 
   @Test
   public void testStorageInvalid7() {
     try {
-      String strStorage = "[f1, f2] serialize by xyz compress by gzip; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo2";
+      String strStorage = "[f1, f2] serialize by xyz compress by gzip; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo";
       Partition p = new Partition(schema.toString(), strStorage);
       CGSchema[] cgschemas = p.getCGSchemas();
       CGSchema cgs1 = cgschemas[0];
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Encountered \" \"serialize by\" \"serialize by \"\" at line 1, column 10.";
+      String str = "Encountered \" <IDENTIFIER> \"xyz \"\" at line 1, column 23.";
       System.out.println(errMsg);
       System.out.println(str);
       Assert.assertEquals(errMsg.startsWith(str), true);
@@ -316,17 +316,17 @@
   @Test
   public void testStorageInvalid8() {
     try {
-      String strStorage = "[f1, f2] serialize by avro compress by xyz; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo2";
+      String strStorage = "[f1, f2] serialize by avro compress by xyz; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo";
       Partition p = new Partition(schema.toString(), strStorage);
       CGSchema[] cgschemas = p.getCGSchemas();
       CGSchema cgs1 = cgschemas[0];
       System.out.println(cgs1);
     } catch (Exception e) {
       String errMsg = e.getMessage();
-      String str = "Encountered \" \"compress by\" \"compress by \"\" at line 1, column 28.";
+      String str = "Encountered \" <IDENTIFIER> \"xyz \"\" at line 1, column 40.";
       System.out.println(errMsg);
       System.out.println(str);
       Assert.assertEquals(errMsg.startsWith(str), true);
     }
   }
-}
\ No newline at end of file
+}