You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pig.apache.org by ra...@apache.org on 2009/10/09 07:04:11 UTC
svn commit: r823411 [2/2] - in /hadoop/pig/trunk/contrib/zebra: ./
src/java/org/apache/hadoop/zebra/io/ src/java/org/apache/hadoop/zebra/types/
src/test/org/apache/hadoop/zebra/io/ src/test/org/apache/hadoop/zebra/pig/
src/test/org/apache/hadoop/zebra/...
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java Fri Oct 9 05:04:10 2009
@@ -69,8 +69,8 @@
public void testInsertNullValues() throws IOException, ParseException {
fs.delete(path, true);
System.out.println("testInsertNullValues");
- writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part1", true);
// Tuple row = TypesUtils.createTuple(writer.getSchema());
// ins.insert(new BytesWritable("key".getBytes()), row);
@@ -82,8 +82,8 @@
public void testFailureInvalidSchema() throws IOException, ParseException {
fs.delete(path, true);
System.out.println("testFailureInvalidSchema");
- writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part1", true);
Tuple row = TypesUtils.createTuple(Schema.parse("xyz, ijk, def"));
try {
@@ -102,8 +102,8 @@
ParseException {
fs.delete(path, true);
System.out.println("testFailureGetInserterAfterWriterClosed");
- writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
try {
writer.close();
TableInserter ins = writer.getInserter("part1", true);
@@ -120,8 +120,8 @@
ParseException {
fs.delete(path, true);
System.out.println("testFailureInsertAfterClose");
- writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part1", true);
Tuple row = TypesUtils.createTuple(writer.getSchema());
@@ -151,7 +151,7 @@
ParseException {
fs.delete(path, true);
System.out.println("testFailureInsertXtraColumn");
- writer = new ColumnGroup.Writer(path, "abc ", false, "pig", "lzo2", true,
+ writer = new ColumnGroup.Writer(path, "abc ", false, "pig", "gz", null, null, (short) -1, true,
conf);
TableInserter ins = writer.getInserter("part1", true);
@@ -164,7 +164,7 @@
row.set(1, map);
Assert
.fail("Failed to catch insertion an extra column not defined in schema.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// noop, expecting exceptions
} finally {
ins.close();
@@ -177,8 +177,8 @@
ParseException {
fs.delete(path, true);
System.out.println("testInsertOneRow");
- writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part1", true);
Tuple row = TypesUtils.createTuple(writer.getSchema());
@@ -199,8 +199,8 @@
ParseException {
fs.delete(path, true);
System.out.println("testInsert2Rows");
- writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part1", true);
// row 1
@@ -233,8 +233,8 @@
ParseException {
fs.delete(path, true);
System.out.println("testInsert2Inserters");
- writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins1 = writer.getInserter("part1", true);
TableInserter ins2 = writer.getInserter("part2", true);
@@ -270,8 +270,8 @@
ParseException {
fs.delete(path, true);
System.out.println("testFailureOverlappingKeys");
- writer = new ColumnGroup.Writer(path, "abc, def ", true, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def ", true, "pig", "gz",
+ null, null, (short) -1, true, conf);
TableInserter ins1 = writer.getInserter("part1", false);
TableInserter ins2 = writer.getInserter("part2", false);
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java Fri Oct 9 05:04:10 2009
@@ -62,8 +62,8 @@
@Test
public void testNew() throws IOException, ParseException {
System.out.println("testNew");
- writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
// NOTE: don't call writer.close() here
close();
}
@@ -73,11 +73,11 @@
ParseException {
System.out.println("testFailureExistingSortedDiff");
try {
- writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz",
+ null, null, (short) -1, true, conf);
finish();
- writer = new ColumnGroup.Writer(path, "abc, def", true, "pig", "lzo2",
- false, conf);
+ writer = new ColumnGroup.Writer(path, "abc, def", true, "pig", "gz",
+ null, null, (short) -1, false, conf);
Assert.fail("Failed to catch sorted flag alteration.");
} catch (IOException e) {
// noop, expecting exceptions
@@ -89,8 +89,8 @@
@Test
public void testExisting() throws IOException, ParseException {
System.out.println("testExisting");
- writer = new ColumnGroup.Writer(path, " abc , def ", false, "pig", "lzo2",
- false, conf);
+ writer = new ColumnGroup.Writer(path, " abc , def ", false, "pig", "gz",
+ null, null, (short) -1, false, conf);
writer.close();
close();
}
@@ -105,7 +105,7 @@
FSDataOutputStream in = fs.create(path);
in.close();
writer = new ColumnGroup.Writer(path, " abc , def ", false, "pig",
- "lzo2", false, conf);
+ "gz", null, null, (short) -1, false, conf);
Assert.fail("Failed to catch path not a directory.");
} catch (IOException e) {
// noop, expecting exceptions
@@ -121,7 +121,7 @@
fs.delete(path, true);
FSDataOutputStream in = fs.create(new Path(path, ColumnGroup.META_FILE));
in.close();
- writer = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", false,
+ writer = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, false,
conf);
Assert.fail("Failed to catch meta file existence.");
} catch (IOException e) {
@@ -135,10 +135,10 @@
public void testFailureDiffSchema() throws IOException, ParseException {
System.out.println("testFailureDiffSchema");
try {
- writer = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", false,
+ writer = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, false,
conf);
writer.finish();
- writer = new ColumnGroup.Writer(path, "efg", false, "pig", "lzo2", false,
+ writer = new ColumnGroup.Writer(path, "efg", false, "pig", "gz", null, null, (short) -1, false,
conf);
Assert.fail("Failed to catch schema differences.");
} catch (IOException e) {
@@ -155,7 +155,7 @@
ColumnGroup.Writer writer2 = null;
ColumnGroup.Writer writer3 = null;
try {
- writer1 = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", true,
+ writer1 = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, true,
conf);
writer2 = new ColumnGroup.Writer(path, conf);
writer3 = new ColumnGroup.Writer(path, conf);
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java Fri Oct 9 05:04:10 2009
@@ -63,7 +63,7 @@
schema = new Schema("a,b,c,d,e,f,g");
ColumnGroup.Writer writer = new ColumnGroup.Writer(path, schema, false,
- "pig", "lzo2", true, conf);
+ "pig", "gz", null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part0", true);
// row 1
@@ -168,7 +168,7 @@
try {
row.get(0);
Assert.fail("Failed to catch out of boundary exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -202,7 +202,7 @@
try {
row.get(1);
Assert.fail("Failed to catch out of boundary exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -236,7 +236,7 @@
try {
row.get(1);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -270,7 +270,7 @@
try {
row.get(2);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -304,7 +304,7 @@
try {
row.get(2);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java Fri Oct 9 05:04:10 2009
@@ -77,7 +77,7 @@
System.out.println("testInsert2Inserters");
boolean sorted = false; // true;
writer = new ColumnGroup.Writer(path, "col1, colTWO", sorted, "pig",
- "lzo2", true, conf);
+ "gz", null, null, (short) -1, true, conf);
TableInserter ins1 = writer.getInserter("part1", false);
TableInserter ins2 = writer.getInserter("part2", false);
@@ -197,8 +197,8 @@
throws IOException, ExecException, ParseException {
ColumnGroup.Writer writer = null;
if (schemaString != null) {
- writer = new ColumnGroup.Writer(path, schemaString, false, "pig", "lzo2",
- true, conf);
+ writer = new ColumnGroup.Writer(path, schemaString, false, "pig", "gz",
+ null, null, (short) -1, true, conf);
} else {
writer = new ColumnGroup.Writer(path, conf);
}
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java Fri Oct 9 05:04:10 2009
@@ -73,7 +73,7 @@
schema = new Schema(STR_SCHEMA);
ColumnGroup.Writer writer = new ColumnGroup.Writer(path, schema, false,
- "pig", "lzo2", true, conf);
+ "pig", "gz", null, null, (short) -1, true, conf);
TableInserter ins = writer.getInserter("part0", true);
Tuple row = TypesUtils.createTuple(schema);
@@ -273,7 +273,7 @@
try {
row.get(0);
Assert.fail("Failed to catch out of boundary exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -306,7 +306,7 @@
try {
row.get(1);
Assert.fail("Failed to catch out of boundary exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -340,7 +340,7 @@
try {
row.get(1);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -374,7 +374,7 @@
try {
row.get(2);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -415,7 +415,7 @@
try {
Object obj = tupColl.get(1);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
tupColl = itorColl.next();
@@ -424,7 +424,7 @@
try {
row.get(3);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -474,7 +474,7 @@
try {
Object obj = tupColl.get(1);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
tupColl = itorColl.next();
@@ -482,7 +482,7 @@
try {
Object obj = tupColl.get(1);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
Assert.assertFalse(itorColl.hasNext());
@@ -492,7 +492,7 @@
try {
Object obj = row.get(4);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
@@ -545,7 +545,7 @@
try {
row.get(3);
Assert.fail("Failed to catch 'out of boundary' exceptions.");
- } catch (ExecException e) {
+ } catch (IndexOutOfBoundsException e) {
// no op, expecting out of bounds exceptions
}
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java Fri Oct 9 05:04:10 2009
@@ -210,7 +210,7 @@
@Test
public void testRead1() throws IOException, ParseException {
String query = "records = LOAD '" + path.toString()
- + "' USING org.apache.pig.table.pig.TableLoader('c');";
+ + "' USING org.apache.hadoop.zebra.pig.TableLoader('c');";
System.out.println(query);
pigServer.registerQuery(query);
Iterator<Tuple> it = pigServer.openIterator("records");
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java Fri Oct 9 05:04:10 2009
@@ -262,7 +262,7 @@
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Encountered \" \"[\" \"[ \"\" at line 1, column 1.";
+ String str = "Encountered \" \"{\" \"{ \"\" at line 1, column 4.";
System.out.println(errMsg);
System.out.println(str);
Assert.assertEquals(errMsg.startsWith(str), true);
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java Fri Oct 9 05:04:10 2009
@@ -81,8 +81,15 @@
@Test
public void testStorageValid1() {
try {
- String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip ; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo2 SERIALIZE BY avro";
- Partition p = new Partition(schema.toString(), strStorage);
+ String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav group:data perm:0766; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav group:data perm:966; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:ggg SECURE BY group:fff; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gauravj user:gauravj; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro";
+
+ Partition p = new Partition(schema.toString(), strStorage);
CGSchema[] cgschemas = p.getCGSchemas();
// 3 column group;
@@ -157,4 +164,4 @@
Assert.assertTrue(false);
}
}
-}
\ No newline at end of file
+}
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java Fri Oct 9 05:04:10 2009
@@ -207,9 +207,9 @@
Assert.assertEquals(cgs1.getCompressor(), "gzip");
Assert.assertEquals(cgs1.getSerializer(), "pig");
- Assert.assertEquals(cgs2.getCompressor(), "lzo2");
+ Assert.assertEquals(cgs2.getCompressor(), "lzo");
Assert.assertEquals(cgs2.getSerializer(), "avro");
- Assert.assertEquals(cgs3.getCompressor(), "lzo2");
+ Assert.assertEquals(cgs3.getCompressor(), "lzo");
Assert.assertEquals(cgs3.getSerializer(), "pig");
System.out.println("*********** Column Map **********");
@@ -249,4 +249,4 @@
} catch (Exception e) {
}
}
-}
\ No newline at end of file
+}
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java
URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java?rev=823411&r1=823410&r2=823411&view=diff
==============================================================================
--- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java (original)
+++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java Fri Oct 9 05:04:10 2009
@@ -107,11 +107,11 @@
Assert.assertEquals("f6", f32.name);
Assert.assertEquals(ColumnType.BYTES, f32.type);
- Assert.assertEquals(cgs1.getCompressor(), "lzo2");
+ Assert.assertEquals(cgs1.getCompressor(), "lzo");
Assert.assertEquals(cgs1.getSerializer(), "pig");
Assert.assertEquals(cgs2.getCompressor(), "gzip");
Assert.assertEquals(cgs2.getSerializer(), "avro");
- Assert.assertEquals(cgs3.getCompressor(), "lzo2");
+ Assert.assertEquals(cgs3.getCompressor(), "lzo");
Assert.assertEquals(cgs3.getSerializer(), "pig");
System.out.println("*********** Column Map **********");
@@ -204,10 +204,10 @@
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n <EOF> \n \"compress by\" ...\n \"serialize by\" ...\n \"[\" ...\n ";
+ String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\n";
System.out.println(errMsg);
System.out.println(str);
- Assert.assertEquals(errMsg, str);
+ Assert.assertEquals(errMsg.startsWith(str), true);
}
}
@@ -238,10 +238,10 @@
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n <EOF> \n \"compress by\" ...\n \"serialize by\" ...\n \"[\" ...\n ";
+ String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n";
System.out.println(errMsg);
System.out.println(str);
- Assert.assertEquals(errMsg, str);
+ Assert.assertEquals(errMsg.startsWith(str), true);
}
}
@@ -255,10 +255,10 @@
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Encountered \" \"[\" \"[ \"\" at line 1, column 1.\nWas expecting one of:\n \"compress by\" ...\n \"serialize by\" ...\n ";
+ String str = "Encountered \"<EOF>\" at line 1, column 1.\nWas expecting one of:\n";
System.out.println(errMsg);
System.out.println(str);
- Assert.assertEquals(errMsg, str);
+ Assert.assertEquals(errMsg.startsWith(str), true);
}
}
@@ -289,24 +289,24 @@
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Lexical error at line 1, column 2. Encountered: <EOF> after : \"\"";
+ String str = "Encountered \" \":\" \": \"\" at line 1, column 1.\n";
System.out.println(errMsg);
System.out.println(str);
- Assert.assertEquals(errMsg, str);
+ Assert.assertEquals(errMsg.startsWith(str), true);
}
}
@Test
public void testStorageInvalid7() {
try {
- String strStorage = "[f1, f2] serialize by xyz compress by gzip; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo2";
+ String strStorage = "[f1, f2] serialize by xyz compress by gzip; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo";
Partition p = new Partition(schema.toString(), strStorage);
CGSchema[] cgschemas = p.getCGSchemas();
CGSchema cgs1 = cgschemas[0];
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Encountered \" \"serialize by\" \"serialize by \"\" at line 1, column 10.";
+ String str = "Encountered \" <IDENTIFIER> \"xyz \"\" at line 1, column 23.";
System.out.println(errMsg);
System.out.println(str);
Assert.assertEquals(errMsg.startsWith(str), true);
@@ -316,17 +316,17 @@
@Test
public void testStorageInvalid8() {
try {
- String strStorage = "[f1, f2] serialize by avro compress by xyz; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo2";
+ String strStorage = "[f1, f2] serialize by avro compress by xyz; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo";
Partition p = new Partition(schema.toString(), strStorage);
CGSchema[] cgschemas = p.getCGSchemas();
CGSchema cgs1 = cgschemas[0];
System.out.println(cgs1);
} catch (Exception e) {
String errMsg = e.getMessage();
- String str = "Encountered \" \"compress by\" \"compress by \"\" at line 1, column 28.";
+ String str = "Encountered \" <IDENTIFIER> \"xyz \"\" at line 1, column 40.";
System.out.println(errMsg);
System.out.println(str);
Assert.assertEquals(errMsg.startsWith(str), true);
}
}
-}
\ No newline at end of file
+}