You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@drill.apache.org by pa...@apache.org on 2015/02/14 06:51:40 UTC

[2/5] drill git commit: DRILL-1948: Fix issue with HDFS and CompatibilityUtil.getBuf

DRILL-1948: Fix issue with HDFS and CompatibilityUtil.getBuf


Project: http://git-wip-us.apache.org/repos/asf/drill/repo
Commit: http://git-wip-us.apache.org/repos/asf/drill/commit/30769783
Tree: http://git-wip-us.apache.org/repos/asf/drill/tree/30769783
Diff: http://git-wip-us.apache.org/repos/asf/drill/diff/30769783

Branch: refs/heads/master
Commit: 30769783ec22503d4ab0265da0b449df46a856fd
Parents: 02d51dd
Author: AdamPD <ad...@pharmadata.net.au>
Authored: Mon Feb 9 11:54:54 2015 +1000
Committer: Parth Chandra <pc...@maprtech.com>
Committed: Fri Feb 13 21:50:29 2015 -0800

----------------------------------------------------------------------
 .../src/main/java/parquet/hadoop/ColumnChunkIncReadStore.java    | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/drill/blob/30769783/exec/java-exec/src/main/java/parquet/hadoop/ColumnChunkIncReadStore.java
----------------------------------------------------------------------
diff --git a/exec/java-exec/src/main/java/parquet/hadoop/ColumnChunkIncReadStore.java b/exec/java-exec/src/main/java/parquet/hadoop/ColumnChunkIncReadStore.java
index 2a3124c..f2fe376 100644
--- a/exec/java-exec/src/main/java/parquet/hadoop/ColumnChunkIncReadStore.java
+++ b/exec/java-exec/src/main/java/parquet/hadoop/ColumnChunkIncReadStore.java
@@ -144,7 +144,9 @@ public class ColumnChunkIncReadStore implements PageReadStore {
               ByteBuf buf = allocator.buffer(pageHeader.compressed_page_size);
               lastPage = buf;
               ByteBuffer buffer = buf.nioBuffer(0, pageHeader.compressed_page_size);
-              CompatibilityUtil.getBuf(in, buffer, pageHeader.compressed_page_size);
+              while (buffer.remaining() > 0) {
+                CompatibilityUtil.getBuf(in, buffer, pageHeader.compressed_page_size);
+              }
               return new Page(
                       decompressor.decompress(BytesInput.from(buffer, 0, pageHeader.compressed_page_size), pageHeader.getUncompressed_page_size()),
                       pageHeader.data_page_header.num_values,