You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by pr...@apache.org on 2014/06/17 19:59:26 UTC
svn commit: r1603240 [2/2] - in /hive/trunk:
common/src/java/org/apache/hadoop/hive/conf/ conf/
ql/src/java/org/apache/hadoop/hive/ql/io/orc/
ql/src/test/org/apache/hadoop/hive/ql/io/orc/
ql/src/test/results/clientpositive/ ql/src/test/results/clientpo...
Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestNewIntegerEncoding.java Tue Jun 17 17:59:25 2014
@@ -21,12 +21,15 @@ import static junit.framework.Assert.ass
import java.io.File;
import java.sql.Timestamp;
+import java.util.Arrays;
+import java.util.Collection;
import java.util.List;
import java.util.Random;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hive.ql.io.orc.OrcFile.EncodingStrategy;
import org.apache.hadoop.hive.serde2.io.TimestampWritable;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
@@ -36,12 +39,29 @@ import org.junit.Before;
import org.junit.Rule;
import org.junit.Test;
import org.junit.rules.TestName;
+import org.junit.runner.RunWith;
+import org.junit.runners.Parameterized;
+import org.junit.runners.Parameterized.Parameters;
import com.google.common.collect.Lists;
import com.google.common.primitives.Longs;
+@RunWith(value = Parameterized.class)
public class TestNewIntegerEncoding {
+ private EncodingStrategy encodingStrategy;
+
+ public TestNewIntegerEncoding(EncodingStrategy es) {
+ this.encodingStrategy = es;
+ }
+
+ @Parameters
+ public static Collection<Object[]> data() {
+ Object[][] data = new Object[][] { { EncodingStrategy.COMPRESSION },
+ { EncodingStrategy.SPEED } };
+ return Arrays.asList(data);
+ }
+
public static class TSRow {
Timestamp ts;
@@ -92,7 +112,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
writer.addRow(new Row(111, 1111L));
writer.addRow(new Row(111, 1111L));
writer.addRow(new Row(111, 1111L));
@@ -127,7 +148,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.compress(CompressionKind.NONE)
.version(OrcFile.Version.V_0_11)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -163,7 +185,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -195,7 +218,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -227,7 +251,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -259,7 +284,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -291,7 +317,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -322,7 +349,8 @@ public class TestNewIntegerEncoding {
OrcFile.writerOptions(conf)
.inspector(inspector)
.stripeSize(100000)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -354,7 +382,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -386,7 +415,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -418,7 +448,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -453,7 +484,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -488,7 +520,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -531,7 +564,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -574,7 +608,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -617,7 +652,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -651,7 +687,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -687,7 +724,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -723,7 +761,8 @@ public class TestNewIntegerEncoding {
.inspector(inspector)
.stripeSize(100000)
.compress(CompressionKind.NONE)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -758,7 +797,8 @@ public class TestNewIntegerEncoding {
OrcFile.writerOptions(conf)
.inspector(inspector)
.stripeSize(100000)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -793,7 +833,8 @@ public class TestNewIntegerEncoding {
OrcFile.writerOptions(conf)
.inspector(inspector)
.stripeSize(100000)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -828,7 +869,8 @@ public class TestNewIntegerEncoding {
OrcFile.writerOptions(conf)
.inspector(inspector)
.stripeSize(100000)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -863,7 +905,8 @@ public class TestNewIntegerEncoding {
OrcFile.writerOptions(conf)
.inspector(inspector)
.stripeSize(100000)
- .bufferSize(10000));
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
@@ -895,7 +938,11 @@ public class TestNewIntegerEncoding {
input.set(511, Long.MAX_VALUE);
Writer writer = OrcFile.createWriter(testFilePath,
- OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+ OrcFile.writerOptions(conf)
+ .inspector(inspector)
+ .stripeSize(100000)
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for (Long l : input) {
writer.addRow(l);
}
@@ -929,7 +976,11 @@ public class TestNewIntegerEncoding {
input.set(511, Long.MAX_VALUE);
Writer writer = OrcFile.createWriter(testFilePath,
- OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+ OrcFile.writerOptions(conf)
+ .inspector(inspector)
+ .stripeSize(100000)
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for (Long l : input) {
writer.addRow(l);
}
@@ -976,7 +1027,11 @@ public class TestNewIntegerEncoding {
input.add(33333L);
Writer writer = OrcFile.createWriter(testFilePath,
- OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+ OrcFile.writerOptions(conf)
+ .inspector(inspector)
+ .stripeSize(100000)
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for (Long l : input) {
writer.addRow(l);
}
@@ -1026,7 +1081,11 @@ public class TestNewIntegerEncoding {
input.add(Long.MAX_VALUE);
Writer writer = OrcFile.createWriter(testFilePath,
- OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+ OrcFile.writerOptions(conf)
+ .inspector(inspector)
+ .stripeSize(100000)
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
for (Long l : input) {
writer.addRow(l);
}
@@ -1051,7 +1110,11 @@ public class TestNewIntegerEncoding {
}
Writer writer = OrcFile.createWriter(testFilePath,
- OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+ OrcFile.writerOptions(conf)
+ .inspector(inspector)
+ .stripeSize(100000)
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
List<Timestamp> tslist = Lists.newArrayList();
tslist.add(Timestamp.valueOf("9999-01-01 00:00:00"));
@@ -1114,7 +1177,11 @@ public class TestNewIntegerEncoding {
}
Writer writer = OrcFile.createWriter(testFilePath,
- OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000).bufferSize(10000));
+ OrcFile.writerOptions(conf)
+ .inspector(inspector)
+ .stripeSize(100000)
+ .bufferSize(10000)
+ .encodingStrategy(encodingStrategy));
writer.addRow(-7486502418706614742L);
writer.addRow(0L);
@@ -1157,7 +1224,8 @@ public class TestNewIntegerEncoding {
.compress(CompressionKind.NONE)
.stripeSize(100000)
.bufferSize(10000)
- .version(OrcFile.Version.V_0_11));
+ .version(OrcFile.Version.V_0_11)
+ .encodingStrategy(encodingStrategy));
for(Long l : input) {
writer.addRow(l);
}
Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestSerializationUtils.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestSerializationUtils.java?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestSerializationUtils.java (original)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestSerializationUtils.java Tue Jun 17 17:59:25 2014
@@ -35,10 +35,15 @@ public class TestSerializationUtils {
@Test
public void testDoubles() throws Exception {
+ double tolerance = 0.0000000000000001;
ByteArrayOutputStream buffer = new ByteArrayOutputStream();
- SerializationUtils.writeDouble(buffer, 1343822337.759);
- assertEquals(1343822337.759,
- SerializationUtils.readDouble(fromBuffer(buffer)), 0.0001);
+ SerializationUtils utils = new SerializationUtils();
+ utils.writeDouble(buffer, 1343822337.759);
+ assertEquals(1343822337.759, utils.readDouble(fromBuffer(buffer)), tolerance);
+ buffer = new ByteArrayOutputStream();
+ utils.writeDouble(buffer, 0.8);
+ double got = utils.readDouble(fromBuffer(buffer));
+ assertEquals(0.8, got, tolerance);
}
@Test
Added: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestUnrolledBitPack.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestUnrolledBitPack.java?rev=1603240&view=auto
==============================================================================
--- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestUnrolledBitPack.java (added)
+++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestUnrolledBitPack.java Tue Jun 17 17:59:25 2014
@@ -0,0 +1,95 @@
+package org.apache.hadoop.hive.ql.io.orc;
+
+import static org.junit.Assert.assertEquals;
+
+import java.io.File;
+import java.util.Arrays;
+import java.util.Collection;
+import java.util.List;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
+import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory;
+import org.apache.hadoop.io.LongWritable;
+import org.junit.Before;
+import org.junit.Rule;
+import org.junit.Test;
+import org.junit.rules.TestName;
+import org.junit.runner.RunWith;
+import org.junit.runners.Parameterized;
+import org.junit.runners.Parameterized.Parameters;
+
+import com.google.common.collect.Lists;
+import com.google.common.primitives.Longs;
+
+@RunWith(value = Parameterized.class)
+public class TestUnrolledBitPack {
+
+ private long val;
+
+ public TestUnrolledBitPack(long val) {
+ this.val = val;
+ }
+
+ @Parameters
+ public static Collection<Object[]> data() {
+ Object[][] data = new Object[][] { { -1 }, { 1 }, { 7 }, { -128 }, { 32000 }, { 8300000 },
+ { Integer.MAX_VALUE }, { 540000000000L }, { 140000000000000L }, { 36000000000000000L },
+ { Long.MAX_VALUE } };
+ return Arrays.asList(data);
+ }
+
+ Path workDir = new Path(System.getProperty("test.tmp.dir", "target" + File.separator + "test"
+ + File.separator + "tmp"));
+
+ Configuration conf;
+ FileSystem fs;
+ Path testFilePath;
+
+ @Rule
+ public TestName testCaseName = new TestName();
+
+ @Before
+ public void openFileSystem() throws Exception {
+ conf = new Configuration();
+ fs = FileSystem.getLocal(conf);
+ testFilePath = new Path(workDir, "TestOrcFile." + testCaseName.getMethodName() + ".orc");
+ fs.delete(testFilePath, false);
+ }
+
+ @Test
+ public void testBitPacking() throws Exception {
+ ObjectInspector inspector;
+ synchronized (TestOrcFile.class) {
+ inspector = ObjectInspectorFactory.getReflectionObjectInspector(Long.class,
+ ObjectInspectorFactory.ObjectInspectorOptions.JAVA);
+ }
+
+ long[] inp = new long[] { val, 0, val, val, 0, val, 0, val, val, 0, val, 0, val, val, 0, 0,
+ val, val, 0, val, 0, 0, val, 0, val, 0, val, 0, 0, val, 0, val, 0, val, 0, 0, val, 0, val,
+ 0, val, 0, 0, val, 0, val, 0, val, 0, 0, val, 0, val, 0, val, 0, 0, val, 0, val, 0, val, 0,
+ 0, val, 0, val, 0, val, 0, 0, val, 0, val, 0, val, 0, 0, val, 0, val, 0, val, 0, 0, val, 0,
+ val, 0, val, 0, 0, val, 0, val, 0, 0, val, val };
+ List<Long> input = Lists.newArrayList(Longs.asList(inp));
+
+ Writer writer = OrcFile.createWriter(
+ testFilePath,
+ OrcFile.writerOptions(conf).inspector(inspector).stripeSize(100000)
+ .compress(CompressionKind.NONE).bufferSize(10000));
+ for (Long l : input) {
+ writer.addRow(l);
+ }
+ writer.close();
+
+ Reader reader = OrcFile.createReader(testFilePath, OrcFile.readerOptions(conf).filesystem(fs));
+ RecordReader rows = reader.rows();
+ int idx = 0;
+ while (rows.hasNext()) {
+ Object row = rows.next(null);
+ assertEquals(input.get(idx++).longValue(), ((LongWritable) row).get());
+ }
+ }
+
+}
Modified: hive/trunk/ql/src/test/results/clientpositive/annotate_stats_filter.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/annotate_stats_filter.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
Files hive/trunk/ql/src/test/results/clientpositive/annotate_stats_filter.q.out (original) and hive/trunk/ql/src/test/results/clientpositive/annotate_stats_filter.q.out Tue Jun 17 17:59:25 2014 differ
Modified: hive/trunk/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
Files hive/trunk/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out (original) and hive/trunk/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out Tue Jun 17 17:59:25 2014 differ
Modified: hive/trunk/ql/src/test/results/clientpositive/annotate_stats_join.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/annotate_stats_join.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
Files hive/trunk/ql/src/test/results/clientpositive/annotate_stats_join.q.out (original) and hive/trunk/ql/src/test/results/clientpositive/annotate_stats_join.q.out Tue Jun 17 17:59:25 2014 differ
Modified: hive/trunk/ql/src/test/results/clientpositive/annotate_stats_part.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/annotate_stats_part.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
Files hive/trunk/ql/src/test/results/clientpositive/annotate_stats_part.q.out (original) and hive/trunk/ql/src/test/results/clientpositive/annotate_stats_part.q.out Tue Jun 17 17:59:25 2014 differ
Modified: hive/trunk/ql/src/test/results/clientpositive/annotate_stats_union.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/annotate_stats_union.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
Files hive/trunk/ql/src/test/results/clientpositive/annotate_stats_union.q.out (original) and hive/trunk/ql/src/test/results/clientpositive/annotate_stats_union.q.out Tue Jun 17 17:59:25 2014 differ
Modified: hive/trunk/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out Tue Jun 17 17:59:25 2014
@@ -874,7 +874,7 @@ Partition Parameters:
numFiles 2
numRows 16
rawDataSize 320
- totalSize 1318
+ totalSize 1348
#### A masked pattern was here ####
# Storage Information
@@ -918,7 +918,7 @@ Partition Parameters:
numFiles 2
numRows 3
rawDataSize 60
- totalSize 1026
+ totalSize 1034
#### A masked pattern was here ####
# Storage Information
@@ -962,7 +962,7 @@ Partition Parameters:
numFiles 2
numRows 7
rawDataSize 140
- totalSize 1146
+ totalSize 1166
#### A masked pattern was here ####
# Storage Information
@@ -1006,7 +1006,7 @@ Partition Parameters:
numFiles 2
numRows 3
rawDataSize 60
- totalSize 1038
+ totalSize 1040
#### A masked pattern was here ####
# Storage Information
@@ -1049,7 +1049,7 @@ Partition Parameters:
numFiles 8
numRows 16
rawDataSize 320
- totalSize 4294
+ totalSize 4332
#### A masked pattern was here ####
# Storage Information
@@ -1135,7 +1135,7 @@ Partition Parameters:
numFiles 8
numRows 16
rawDataSize 320
- totalSize 4278
+ totalSize 4318
#### A masked pattern was here ####
# Storage Information
Modified: hive/trunk/ql/src/test/results/clientpositive/orc_analyze.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/orc_analyze.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/orc_analyze.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/orc_analyze.q.out Tue Jun 17 17:59:25 2014
@@ -103,7 +103,7 @@ Table Parameters:
numFiles 1
numRows 100
rawDataSize 52600
- totalSize 3033
+ totalSize 3042
#### A masked pattern was here ####
# Storage Information
@@ -192,7 +192,7 @@ Table Parameters:
numFiles 1
numRows 100
rawDataSize 52600
- totalSize 3033
+ totalSize 3042
#### A masked pattern was here ####
# Storage Information
@@ -305,7 +305,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 21950
- totalSize 1948
+ totalSize 1962
#### A masked pattern was here ####
# Storage Information
@@ -350,7 +350,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 22050
- totalSize 1969
+ totalSize 1981
#### A masked pattern was here ####
# Storage Information
@@ -451,7 +451,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 21950
- totalSize 1948
+ totalSize 1962
#### A masked pattern was here ####
# Storage Information
@@ -496,7 +496,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 22050
- totalSize 1969
+ totalSize 1981
#### A masked pattern was here ####
# Storage Information
@@ -615,7 +615,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 21980
- totalSize 4780
+ totalSize 4746
#### A masked pattern was here ####
# Storage Information
@@ -660,7 +660,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 22048
- totalSize 4859
+ totalSize 4829
#### A masked pattern was here ####
# Storage Information
@@ -767,7 +767,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 21980
- totalSize 4780
+ totalSize 4746
#### A masked pattern was here ####
# Storage Information
@@ -812,7 +812,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 22048
- totalSize 4859
+ totalSize 4829
#### A masked pattern was here ####
# Storage Information
@@ -976,7 +976,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 21950
- totalSize 1948
+ totalSize 1962
#### A masked pattern was here ####
# Storage Information
Modified: hive/trunk/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out Tue Jun 17 17:59:25 2014
@@ -914,7 +914,7 @@ Partition Parameters:
numFiles 2
numRows 16
rawDataSize 320
- totalSize 1318
+ totalSize 1348
#### A masked pattern was here ####
# Storage Information
@@ -958,7 +958,7 @@ Partition Parameters:
numFiles 2
numRows 3
rawDataSize 60
- totalSize 1026
+ totalSize 1034
#### A masked pattern was here ####
# Storage Information
@@ -1002,7 +1002,7 @@ Partition Parameters:
numFiles 2
numRows 7
rawDataSize 140
- totalSize 1146
+ totalSize 1166
#### A masked pattern was here ####
# Storage Information
@@ -1046,7 +1046,7 @@ Partition Parameters:
numFiles 2
numRows 3
rawDataSize 60
- totalSize 1038
+ totalSize 1040
#### A masked pattern was here ####
# Storage Information
@@ -1089,7 +1089,7 @@ Partition Parameters:
numFiles 8
numRows 16
rawDataSize 320
- totalSize 4294
+ totalSize 4332
#### A masked pattern was here ####
# Storage Information
@@ -1175,7 +1175,7 @@ Partition Parameters:
numFiles 8
numRows 16
rawDataSize 320
- totalSize 4278
+ totalSize 4318
#### A masked pattern was here ####
# Storage Information
Modified: hive/trunk/ql/src/test/results/clientpositive/tez/orc_analyze.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/tez/orc_analyze.q.out?rev=1603240&r1=1603239&r2=1603240&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/tez/orc_analyze.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/tez/orc_analyze.q.out Tue Jun 17 17:59:25 2014
@@ -103,7 +103,7 @@ Table Parameters:
numFiles 1
numRows 100
rawDataSize 52600
- totalSize 3033
+ totalSize 3042
#### A masked pattern was here ####
# Storage Information
@@ -192,7 +192,7 @@ Table Parameters:
numFiles 1
numRows 100
rawDataSize 52600
- totalSize 3033
+ totalSize 3042
#### A masked pattern was here ####
# Storage Information
@@ -305,7 +305,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 21950
- totalSize 1948
+ totalSize 1962
#### A masked pattern was here ####
# Storage Information
@@ -350,7 +350,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 22050
- totalSize 1969
+ totalSize 1981
#### A masked pattern was here ####
# Storage Information
@@ -451,7 +451,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 21950
- totalSize 1948
+ totalSize 1962
#### A masked pattern was here ####
# Storage Information
@@ -496,7 +496,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 22050
- totalSize 1969
+ totalSize 1981
#### A masked pattern was here ####
# Storage Information
@@ -615,7 +615,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 21980
- totalSize 4780
+ totalSize 4746
#### A masked pattern was here ####
# Storage Information
@@ -660,7 +660,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 22048
- totalSize 4859
+ totalSize 4829
#### A masked pattern was here ####
# Storage Information
@@ -767,7 +767,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 21980
- totalSize 4780
+ totalSize 4746
#### A masked pattern was here ####
# Storage Information
@@ -812,7 +812,7 @@ Partition Parameters:
numFiles 4
numRows 50
rawDataSize 22048
- totalSize 4859
+ totalSize 4829
#### A masked pattern was here ####
# Storage Information
@@ -976,7 +976,7 @@ Partition Parameters:
numFiles 1
numRows 50
rawDataSize 21950
- totalSize 1948
+ totalSize 1962
#### A masked pattern was here ####
# Storage Information