You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by zs...@apache.org on 2010/01/21 10:52:48 UTC
svn commit: r901625 [6/6] - in /hadoop/hive/trunk: ./
serde/src/java/org/apache/hadoop/hive/serde2/
serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/
serde/src/java/org/apache/hadoop/hive/serde2/columnar/
serde/src/java/org/apache/hadoop/hiv...
Modified: hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestUnionStructObjectInspector.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestUnionStructObjectInspector.java?rev=901625&r1=901624&r2=901625&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestUnionStructObjectInspector.java (original)
+++ hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestUnionStructObjectInspector.java Thu Jan 21 09:52:44 2010
@@ -21,14 +21,13 @@
import java.util.Arrays;
import java.util.List;
+import junit.framework.TestCase;
+
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector.Category;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory;
-import junit.framework.TestCase;
-
public class TestUnionStructObjectInspector extends TestCase {
-
public void testUnionStructObjectInspector() throws Throwable {
try {
ArrayList<String> fieldNames1 = new ArrayList<String>();
@@ -36,49 +35,60 @@
fieldNames1.add("secondString");
fieldNames1.add("thirdBoolean");
ArrayList<ObjectInspector> fieldObjectInspectors1 = new ArrayList<ObjectInspector>();
- fieldObjectInspectors1.add(PrimitiveObjectInspectorFactory.javaIntObjectInspector);
- fieldObjectInspectors1.add(PrimitiveObjectInspectorFactory.javaStringObjectInspector);
- fieldObjectInspectors1.add(PrimitiveObjectInspectorFactory.javaBooleanObjectInspector);
- StandardStructObjectInspector soi1 = ObjectInspectorFactory.getStandardStructObjectInspector(
- fieldNames1, fieldObjectInspectors1);
-
+ fieldObjectInspectors1
+ .add(PrimitiveObjectInspectorFactory.javaIntObjectInspector);
+ fieldObjectInspectors1
+ .add(PrimitiveObjectInspectorFactory.javaStringObjectInspector);
+ fieldObjectInspectors1
+ .add(PrimitiveObjectInspectorFactory.javaBooleanObjectInspector);
+ StandardStructObjectInspector soi1 = ObjectInspectorFactory
+ .getStandardStructObjectInspector(fieldNames1, fieldObjectInspectors1);
+
ArrayList<String> fieldNames2 = new ArrayList<String>();
fieldNames2.add("fourthDouble");
fieldNames2.add("fifthLong");
ArrayList<ObjectInspector> fieldObjectInspectors2 = new ArrayList<ObjectInspector>();
- fieldObjectInspectors2.add(PrimitiveObjectInspectorFactory.javaDoubleObjectInspector);
- fieldObjectInspectors2.add(PrimitiveObjectInspectorFactory.javaLongObjectInspector);
- StandardStructObjectInspector soi2 = ObjectInspectorFactory.getStandardStructObjectInspector(
- fieldNames2, fieldObjectInspectors2);
-
- UnionStructObjectInspector usoi1 = ObjectInspectorFactory.getUnionStructObjectInspector(
- Arrays.asList(new StructObjectInspector[]{soi1, soi2}));
- UnionStructObjectInspector usoi2 = ObjectInspectorFactory.getUnionStructObjectInspector(
- Arrays.asList(new StructObjectInspector[]{soi1, soi2}));
-
+ fieldObjectInspectors2
+ .add(PrimitiveObjectInspectorFactory.javaDoubleObjectInspector);
+ fieldObjectInspectors2
+ .add(PrimitiveObjectInspectorFactory.javaLongObjectInspector);
+ StandardStructObjectInspector soi2 = ObjectInspectorFactory
+ .getStandardStructObjectInspector(fieldNames2, fieldObjectInspectors2);
+
+ UnionStructObjectInspector usoi1 = ObjectInspectorFactory
+ .getUnionStructObjectInspector(Arrays
+ .asList(new StructObjectInspector[] { soi1, soi2 }));
+ UnionStructObjectInspector usoi2 = ObjectInspectorFactory
+ .getUnionStructObjectInspector(Arrays
+ .asList(new StructObjectInspector[] { soi1, soi2 }));
+
assertEquals(usoi1, usoi2);
-
+
// metadata
assertEquals(Category.STRUCT, usoi1.getCategory());
List<? extends StructField> fields = usoi1.getAllStructFieldRefs();
assertEquals(5, fields.size());
- for (int i=0; i<5; i++) {
- if (i<=2) {
- assertEquals(fieldNames1.get(i).toLowerCase(), fields.get(i).getFieldName());
- assertEquals(fieldObjectInspectors1.get(i), fields.get(i).getFieldObjectInspector());
+ for (int i = 0; i < 5; i++) {
+ if (i <= 2) {
+ assertEquals(fieldNames1.get(i).toLowerCase(), fields.get(i)
+ .getFieldName());
+ assertEquals(fieldObjectInspectors1.get(i), fields.get(i)
+ .getFieldObjectInspector());
} else {
- assertEquals(fieldNames2.get(i-3).toLowerCase(), fields.get(i).getFieldName());
- assertEquals(fieldObjectInspectors2.get(i-3), fields.get(i).getFieldObjectInspector());
+ assertEquals(fieldNames2.get(i - 3).toLowerCase(), fields.get(i)
+ .getFieldName());
+ assertEquals(fieldObjectInspectors2.get(i - 3), fields.get(i)
+ .getFieldObjectInspector());
}
}
assertEquals(fields.get(1), usoi1.getStructFieldRef("secondString"));
assertEquals(fields.get(4), usoi1.getStructFieldRef("fifthLong"));
-
+
// null
- for (int i=0; i<5; i++) {
+ for (int i = 0; i < 5; i++) {
assertNull(usoi1.getStructFieldData(null, fields.get(i)));
}
-
+
// real struct
ArrayList<Object> struct1 = new ArrayList<Object>(3);
struct1.add(1);
@@ -93,14 +103,15 @@
ArrayList<Object> all = new ArrayList<Object>(5);
all.addAll(struct1);
all.addAll(struct2);
-
- for (int i=0; i<5; i++) {
- assertEquals(all.get(i), usoi1.getStructFieldData(struct, fields.get(i)));
+
+ for (int i = 0; i < 5; i++) {
+ assertEquals(all.get(i), usoi1
+ .getStructFieldData(struct, fields.get(i)));
}
} catch (Throwable e) {
e.printStackTrace();
throw e;
}
}
-
+
}
Modified: hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/thrift_test/CreateSequenceFile.java
URL: http://svn.apache.org/viewvc/hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/thrift_test/CreateSequenceFile.java?rev=901625&r1=901624&r2=901625&view=diff
==============================================================================
--- hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/thrift_test/CreateSequenceFile.java (original)
+++ hadoop/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/thrift_test/CreateSequenceFile.java Thu Jan 21 09:52:44 2010
@@ -25,11 +25,13 @@
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hive.serde2.ByteStream;
+import org.apache.hadoop.hive.serde2.thrift.test.Complex;
+import org.apache.hadoop.hive.serde2.thrift.test.IntString;
import org.apache.hadoop.io.BytesWritable;
import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Writable;
import org.apache.hadoop.mapred.JobConf;
-
import org.apache.thrift.TBase;
import org.apache.thrift.TException;
import org.apache.thrift.protocol.TBinaryProtocol;
@@ -37,10 +39,6 @@
import org.apache.thrift.protocol.TProtocolFactory;
import org.apache.thrift.transport.TIOStreamTransport;
-import org.apache.hadoop.hive.serde2.ByteStream;
-import org.apache.hadoop.hive.serde2.thrift.test.Complex;
-import org.apache.hadoop.hive.serde2.thrift.test.IntString;
-
public class CreateSequenceFile {
public static void usage() {
@@ -49,18 +47,19 @@
}
public static class ThriftSerializer {
-
+
ByteStream.Output bos;
TProtocol outProtocol;
-
+
public ThriftSerializer() {
bos = new ByteStream.Output();
TIOStreamTransport outTransport = new TIOStreamTransport(bos);
TProtocolFactory outFactory = new TBinaryProtocol.Factory();
outProtocol = outFactory.getProtocol(outTransport);
}
-
+
BytesWritable bw = new BytesWritable();
+
public BytesWritable serialize(TBase base) throws TException {
bos.reset();
base.write(outProtocol);
@@ -68,15 +67,15 @@
return bw;
}
}
-
+
public static void main(String[] args) throws Exception {
// Read parameters
int lines = 10;
- List<String> extraArgs = new ArrayList<String>();
- for(int ai=0; ai<args.length; ai++) {
+ List<String> extraArgs = new ArrayList<String>();
+ for (int ai = 0; ai < args.length; ai++) {
if (args[ai].equals("-line") && ai + 1 < args.length) {
- lines = Integer.parseInt(args[ai+1]);
+ lines = Integer.parseInt(args[ai + 1]);
ai++;
} else {
extraArgs.add(args[ai]);
@@ -85,37 +84,38 @@
if (extraArgs.size() != 1) {
usage();
}
-
+
JobConf conf = new JobConf(CreateSequenceFile.class);
-
+
ThriftSerializer serializer = new ThriftSerializer();
-
+
// Open files
- SequenceFile.Writer writer = new SequenceFile.Writer(FileSystem.get(conf), conf, new Path(extraArgs.get(0)),
- BytesWritable.class, BytesWritable.class);
+ SequenceFile.Writer writer = new SequenceFile.Writer(FileSystem.get(conf),
+ conf, new Path(extraArgs.get(0)), BytesWritable.class,
+ BytesWritable.class);
// write to file
BytesWritable key = new BytesWritable();
-
+
Random rand = new Random(20081215);
-
- for(int i=0; i<lines; i++) {
-
+
+ for (int i = 0; i < lines; i++) {
+
ArrayList<Integer> alist = new ArrayList<Integer>();
- alist.add(i); alist.add(i*2); alist.add(i*3);
+ alist.add(i);
+ alist.add(i * 2);
+ alist.add(i * 3);
ArrayList<String> slist = new ArrayList<String>();
- slist.add("" + i*10); slist.add("" + i*100); slist.add("" + i*1000);
+ slist.add("" + i * 10);
+ slist.add("" + i * 100);
+ slist.add("" + i * 1000);
ArrayList<IntString> islist = new ArrayList<IntString>();
- islist.add(new IntString(i*i, ""+ i*i*i, i));
- HashMap<String,String> hash = new HashMap<String,String>();
+ islist.add(new IntString(i * i, "" + i * i * i, i));
+ HashMap<String, String> hash = new HashMap<String, String>();
hash.put("key_" + i, "value_" + i);
-
- Complex complex = new Complex( rand.nextInt(),
- "record_" + (new Integer(i)).toString(),
- alist,
- slist,
- islist,
- hash);
+
+ Complex complex = new Complex(rand.nextInt(), "record_"
+ + (new Integer(i)).toString(), alist, slist, islist, hash);
Writable value = serializer.serialize(complex);
writer.append(key, value);