You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by om...@apache.org on 2014/04/09 22:11:47 UTC

svn commit: r1586123 - /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcAcidRowReader.java

Author: omalley
Date: Wed Apr  9 20:11:47 2014
New Revision: 1586123

URL: http://svn.apache.org/r1586123
Log:
HIVE-6604. Forgotten file.

Added:
    hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcAcidRowReader.java

Added: hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcAcidRowReader.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcAcidRowReader.java?rev=1586123&view=auto
==============================================================================
--- hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcAcidRowReader.java (added)
+++ hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcAcidRowReader.java Wed Apr  9 20:11:47 2014
@@ -0,0 +1,138 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.io.orc;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedBatchUtil;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatchCtx;
+import org.apache.hadoop.hive.ql.io.AcidInputFormat;
+import org.apache.hadoop.hive.ql.io.RecordIdentifier;
+import org.apache.hadoop.hive.ql.metadata.HiveException;
+import org.apache.hadoop.hive.serde2.SerDeException;
+import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
+import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
+import org.apache.hadoop.io.DataOutputBuffer;
+import org.apache.hadoop.io.NullWritable;
+import org.apache.hadoop.mapred.*;
+
+import java.io.IOException;
+
+/**
+ * Implement a RecordReader that stitches together base and delta files to
+ * support tables and partitions stored in the ACID format. It works by using
+ * the non-vectorized ACID reader and moving the data into a vectorized row
+ * batch.
+ */
+class VectorizedOrcAcidRowReader
+    implements org.apache.hadoop.mapred.RecordReader<NullWritable,
+                                                     VectorizedRowBatch> {
+  private final AcidInputFormat.RowReader<OrcStruct> innerReader;
+  private final RecordIdentifier key;
+  private final OrcStruct value;
+  private final VectorizedRowBatchCtx rowBatchCtx;
+  private final ObjectInspector objectInspector;
+  private boolean needToSetPartition = true;
+  private final DataOutputBuffer buffer = new DataOutputBuffer();
+
+  VectorizedOrcAcidRowReader(AcidInputFormat.RowReader<OrcStruct> inner,
+                             Configuration conf,
+                             FileSplit split) throws IOException {
+    this.innerReader = inner;
+    this.key = inner.createKey();
+    this.rowBatchCtx = new VectorizedRowBatchCtx();
+    this.value = inner.createValue();
+    this.objectInspector = inner.getObjectInspector();
+    try {
+      rowBatchCtx.init(conf, split);
+    } catch (ClassNotFoundException e) {
+      throw new IOException("Failed to initialize context", e);
+    } catch (SerDeException e) {
+      throw new IOException("Failed to initialize context", e);
+    } catch (InstantiationException e) {
+      throw new IOException("Failed to initialize context", e);
+    } catch (IllegalAccessException e) {
+      throw new IOException("Failed to initialize context", e);
+    } catch (HiveException e) {
+      throw new IOException("Failed to initialize context", e);
+    }
+  }
+
+  @Override
+  public boolean next(NullWritable nullWritable,
+                      VectorizedRowBatch vectorizedRowBatch
+                      ) throws IOException {
+    vectorizedRowBatch.reset();
+    buffer.reset();
+    if (!innerReader.next(key, value)) {
+      return false;
+    }
+    if (needToSetPartition) {
+      try {
+        rowBatchCtx.addPartitionColsToBatch(vectorizedRowBatch);
+      } catch (HiveException e) {
+        throw new IOException("Problem adding partition column", e);
+      }
+      needToSetPartition = false;
+    }
+    try {
+      VectorizedBatchUtil.addRowToBatch(value,
+          (StructObjectInspector) objectInspector,
+          vectorizedRowBatch.size++, vectorizedRowBatch, buffer);
+      while (vectorizedRowBatch.size < vectorizedRowBatch.selected.length &&
+          innerReader.next(key, value)) {
+        VectorizedBatchUtil.addRowToBatch(value,
+            (StructObjectInspector) objectInspector,
+            vectorizedRowBatch.size++, vectorizedRowBatch, buffer);
+      }
+    } catch (HiveException he) {
+      throw new IOException("error iterating", he);
+    }
+    return true;
+  }
+
+  @Override
+  public NullWritable createKey() {
+    return NullWritable.get();
+  }
+
+  @Override
+  public VectorizedRowBatch createValue() {
+    try {
+      return rowBatchCtx.createVectorizedRowBatch();
+    } catch (HiveException e) {
+      throw new RuntimeException("Error creating a batch", e);
+    }
+  }
+
+  @Override
+  public long getPos() throws IOException {
+    return innerReader.getPos();
+  }
+
+  @Override
+  public void close() throws IOException {
+    innerReader.close();
+  }
+
+  @Override
+  public float getProgress() throws IOException {
+    return innerReader.getProgress();
+  }
+}