You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@parquet.apache.org by ga...@apache.org on 2023/02/23 06:14:39 UTC
[parquet-mr] branch master updated: PARQUET-2243: Support zstd-jni in DirectCodecFactory (#1027)
This is an automated email from the ASF dual-hosted git repository.
gabor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/parquet-mr.git
The following commit(s) were added to refs/heads/master by this push:
new 65f540779 PARQUET-2243: Support zstd-jni in DirectCodecFactory (#1027)
65f540779 is described below
commit 65f540779d9117f578e9ad2eb99bb64007c1da85
Author: Gabor Szadovszky <ga...@apache.org>
AuthorDate: Thu Feb 23 07:14:32 2023 +0100
PARQUET-2243: Support zstd-jni in DirectCodecFactory (#1027)
---
.../apache/parquet/hadoop/DirectCodecFactory.java | 4 +
.../java/org/apache/parquet/hadoop/DirectZstd.java | 144 +++++++++++++++++++++
.../parquet/hadoop/TestDirectCodecFactory.java | 2 -
3 files changed, 148 insertions(+), 2 deletions(-)
diff --git a/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectCodecFactory.java b/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectCodecFactory.java
index d5f13e286..ef2d5bf89 100644
--- a/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectCodecFactory.java
+++ b/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectCodecFactory.java
@@ -118,6 +118,8 @@ class DirectCodecFactory extends CodecFactory implements AutoCloseable {
} else if (codecName == CompressionCodecName.SNAPPY) {
// avoid using the default Snappy codec since it allocates direct buffers at awkward spots.
return new SnappyCompressor();
+ } else if (codecName == CompressionCodecName.ZSTD) {
+ return DirectZstd.createCompressor(configuration, pageSize);
} else {
// todo: create class similar to the SnappyCompressor for zlib and exclude it as
// snappy is above since it also generates allocateDirect calls.
@@ -132,6 +134,8 @@ class DirectCodecFactory extends CodecFactory implements AutoCloseable {
return new NoopDecompressor();
} else if (codecName == CompressionCodecName.SNAPPY ) {
return new SnappyDecompressor();
+ } else if (codecName == CompressionCodecName.ZSTD) {
+ return DirectZstd.createDecompressor(configuration);
} else if (DirectCodecPool.INSTANCE.codec(codec).supportsDirectDecompression()) {
return new FullDirectDecompressor(codecName);
} else {
diff --git a/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectZstd.java b/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectZstd.java
new file mode 100644
index 000000000..1532e83df
--- /dev/null
+++ b/parquet-hadoop/src/main/java/org/apache/parquet/hadoop/DirectZstd.java
@@ -0,0 +1,144 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ * <p>
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * <p>
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.parquet.hadoop;
+
+import com.github.luben.zstd.BufferPool;
+import com.github.luben.zstd.NoPool;
+import com.github.luben.zstd.RecyclingBufferPool;
+import com.github.luben.zstd.Zstd;
+import com.github.luben.zstd.ZstdOutputStream;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.parquet.bytes.ByteBufferAllocator;
+import org.apache.parquet.bytes.BytesInput;
+import org.apache.parquet.hadoop.codec.ZstdDecompressorStream;
+import org.apache.parquet.hadoop.metadata.CompressionCodecName;
+
+import java.io.ByteArrayOutputStream;
+import java.io.IOException;
+import java.nio.ByteBuffer;
+
+import static org.apache.parquet.hadoop.codec.ZstandardCodec.DEFAULT_PARQUET_COMPRESS_ZSTD_BUFFERPOOL_ENABLED;
+import static org.apache.parquet.hadoop.codec.ZstandardCodec.DEFAULT_PARQUET_COMPRESS_ZSTD_LEVEL;
+import static org.apache.parquet.hadoop.codec.ZstandardCodec.DEFAULTPARQUET_COMPRESS_ZSTD_WORKERS;
+import static org.apache.parquet.hadoop.codec.ZstandardCodec.PARQUET_COMPRESS_ZSTD_BUFFERPOOL_ENABLED;
+import static org.apache.parquet.hadoop.codec.ZstandardCodec.PARQUET_COMPRESS_ZSTD_LEVEL;
+import static org.apache.parquet.hadoop.codec.ZstandardCodec.PARQUET_COMPRESS_ZSTD_WORKERS;
+
+/**
+ * Utility class to support creating compressor and decompressor instances for the ZStandard codec. It is implemented in
+ * a way to work around the codec pools implemented in both parquet-mr and hadoop. These codec pools may result creating
+ * and dereferencing direct byte buffers causing OOM errors in case of many parallel compressor/decompressor instances
+ * are required working on direct memory.
+ *
+ * @see DirectCodecFactory.DirectCodecPool
+ * @see org.apache.hadoop.io.compress.CodecPool
+ */
+class DirectZstd {
+
+ static CodecFactory.BytesCompressor createCompressor(Configuration conf, int pageSize) {
+ return new ZstdCompressor(
+ getPool(conf),
+ conf.getInt(PARQUET_COMPRESS_ZSTD_LEVEL, DEFAULT_PARQUET_COMPRESS_ZSTD_LEVEL),
+ conf.getInt(PARQUET_COMPRESS_ZSTD_WORKERS, DEFAULTPARQUET_COMPRESS_ZSTD_WORKERS),
+ pageSize);
+ }
+
+ static CodecFactory.BytesDecompressor createDecompressor(Configuration conf) {
+ return new ZstdDecompressor(getPool(conf));
+ }
+
+ private static class ZstdCompressor extends CodecFactory.BytesCompressor {
+ private final BufferPool pool;
+ private final int level;
+ private final int workers;
+ private final int pageSize;
+
+ ZstdCompressor(BufferPool pool, int level, int workers, int pageSize) {
+ this.pool = pool;
+ this.level = level;
+ this.workers = workers;
+ this.pageSize = pageSize;
+ }
+
+ @Override
+ public BytesInput compress(BytesInput bytes) throws IOException {
+ // Since BytesInput does not support direct memory this implementation is heap based
+ BytesInputProviderOutputStream stream = new BytesInputProviderOutputStream(pageSize);
+ try (ZstdOutputStream zstdStream = new ZstdOutputStream(stream, pool, level)) {
+ zstdStream.setWorkers(workers);
+ bytes.writeAllTo(zstdStream);
+ }
+ return stream.getBytesInput();
+ }
+
+ @Override
+ public CompressionCodecName getCodecName() {
+ return CompressionCodecName.ZSTD;
+ }
+
+ @Override
+ public void release() {
+ // Nothing to do here since we release resources where we create them
+ }
+ }
+
+ private static class ZstdDecompressor extends CodecFactory.BytesDecompressor {
+ private final BufferPool pool;
+
+ private ZstdDecompressor(BufferPool pool) {
+ this.pool = pool;
+ }
+
+ @Override
+ public BytesInput decompress(BytesInput bytes, int uncompressedSize) throws IOException {
+ // Since BytesInput does not support direct memory this implementation is heap based
+ try (ZstdDecompressorStream decompressorStream = new ZstdDecompressorStream(bytes.toInputStream(), pool)) {
+ // We need to copy the bytes from the input stream, so we can close it here (BytesInput does not support closing)
+ return BytesInput.copy(BytesInput.from(decompressorStream, uncompressedSize));
+ }
+ }
+
+ @Override
+ public void decompress(ByteBuffer input, int compressedSize, ByteBuffer output, int uncompressedSize) throws IOException {
+ Zstd.decompress(output, input);
+ }
+
+ @Override
+ public void release() {
+ // Nothing to do here since we release resources where we create them
+ }
+ }
+
+ private static class BytesInputProviderOutputStream extends ByteArrayOutputStream {
+ BytesInputProviderOutputStream(int initialCapacity) {
+ super(initialCapacity);
+ }
+
+ BytesInput getBytesInput() {
+ return BytesInput.from(buf, 0, count);
+ }
+ }
+
+ private static BufferPool getPool(Configuration conf) {
+ if (conf.getBoolean(PARQUET_COMPRESS_ZSTD_BUFFERPOOL_ENABLED, DEFAULT_PARQUET_COMPRESS_ZSTD_BUFFERPOOL_ENABLED)) {
+ return RecyclingBufferPool.INSTANCE;
+ } else {
+ return NoPool.INSTANCE;
+ }
+ }
+}
diff --git a/parquet-hadoop/src/test/java/org/apache/parquet/hadoop/TestDirectCodecFactory.java b/parquet-hadoop/src/test/java/org/apache/parquet/hadoop/TestDirectCodecFactory.java
index 76e6880e8..8fec515a4 100644
--- a/parquet-hadoop/src/test/java/org/apache/parquet/hadoop/TestDirectCodecFactory.java
+++ b/parquet-hadoop/src/test/java/org/apache/parquet/hadoop/TestDirectCodecFactory.java
@@ -35,7 +35,6 @@ import org.apache.parquet.hadoop.metadata.CompressionCodecName;
import static org.apache.parquet.hadoop.metadata.CompressionCodecName.BROTLI;
import static org.apache.parquet.hadoop.metadata.CompressionCodecName.LZ4;
import static org.apache.parquet.hadoop.metadata.CompressionCodecName.LZO;
-import static org.apache.parquet.hadoop.metadata.CompressionCodecName.ZSTD;
public class TestDirectCodecFactory {
@@ -156,7 +155,6 @@ public class TestDirectCodecFactory {
Set<CompressionCodecName> codecsToSkip = new HashSet<>();
codecsToSkip.add(LZO); // not distributed because it is GPL
codecsToSkip.add(LZ4); // not distributed in the default version of Hadoop
- codecsToSkip.add(ZSTD); // not distributed in the default version of Hadoop
final String arch = System.getProperty("os.arch");
if ("aarch64".equals(arch)) {
// PARQUET-1975 brotli-codec does not have natives for ARM64 architectures