You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@tika.apache.org by ta...@apache.org on 2024/03/23 14:48:52 UTC

(tika) branch TIKA-4220-branch2x created (now ea5ae249b)

This is an automated email from the ASF dual-hosted git repository.

tallison pushed a change to branch TIKA-4220-branch2x
in repository https://gitbox.apache.org/repos/asf/tika.git


      at ea5ae249b TIKA-4220 on branch_2x -- temporary commons-compress workaround

This branch includes the following new commits:

     new ea5ae249b TIKA-4220 on branch_2x -- temporary commons-compress workaround

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



(tika) 01/01: TIKA-4220 on branch_2x -- temporary commons-compress workaround

Posted by ta...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

tallison pushed a commit to branch TIKA-4220-branch2x
in repository https://gitbox.apache.org/repos/asf/tika.git

commit ea5ae249b3a99d48bcf961e11384654de5f7eae3
Author: tallison <ta...@apache.org>
AuthorDate: Sat Mar 23 10:48:35 2024 -0400

    TIKA-4220 on branch_2x -- temporary commons-compress workaround
---
 .../org/apache/tika/parser/pkg/PackageParser.java  |  27 +-
 .../detect/zip/DefaultZipContainerDetector.java    |   5 +-
 .../tika/detect/zip/TikaArchiveStreamFactory.java  | 641 +++++++++++++++++++++
 3 files changed, 663 insertions(+), 10 deletions(-)

diff --git a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
index 86bfbaa88..f45fa23bd 100644
--- a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
+++ b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-pkg-module/src/main/java/org/apache/tika/parser/pkg/PackageParser.java
@@ -61,6 +61,7 @@ import org.xml.sax.helpers.AttributesImpl;
 
 import org.apache.tika.config.Field;
 import org.apache.tika.detect.EncodingDetector;
+import org.apache.tika.detect.zip.TikaArchiveStreamFactory;
 import org.apache.tika.exception.EncryptedDocumentException;
 import org.apache.tika.exception.TikaException;
 import org.apache.tika.extractor.EmbeddedDocumentExtractor;
@@ -256,14 +257,24 @@ public class PackageParser extends AbstractEncodingDetectorParser {
         String encoding = null;
         try {
             ArchiveStreamFactory factory =
-                    context.get(ArchiveStreamFactory.class, new ArchiveStreamFactory());
-            encoding = factory.getEntryEncoding();
-            // At the end we want to close the archive stream to release
-            // any associated resources, but the underlying document stream
-            // should not be closed
-
-            ais = factory.createArchiveInputStream(new CloseShieldInputStream(stream));
-
+                    context.get(ArchiveStreamFactory.class);
+            //TODO -- fix this when we next upgrade commons-compress
+            //TODO -- we've probably already detected the stream by here. We should
+            //rely on that detection and not re-detect.
+            if (factory != null) {
+                encoding = factory.getEntryEncoding();
+                // At the end we want to close the archive stream to release
+                // any associated resources, but the underlying document stream
+                // should not be closed
+                ais = factory.createArchiveInputStream(new CloseShieldInputStream(stream));
+            } else {
+                TikaArchiveStreamFactory tikaFactory = new TikaArchiveStreamFactory();
+                encoding = tikaFactory.getEntryEncoding();
+                // At the end we want to close the archive stream to release
+                // any associated resources, but the underlying document stream
+                // should not be closed
+                ais = tikaFactory.createArchiveInputStream(new CloseShieldInputStream(stream));
+            }
         } catch (StreamingNotSupportedException sne) {
             // Most archive formats work on streams, but a few need files
             if (sne.getFormat().equals(ArchiveStreamFactory.SEVEN_Z)) {
diff --git a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
index d2f28e0a0..9842fc0af 100644
--- a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
+++ b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/DefaultZipContainerDetector.java
@@ -25,7 +25,6 @@ import java.util.ArrayList;
 import java.util.List;
 
 import org.apache.commons.compress.archivers.ArchiveException;
-import org.apache.commons.compress.archivers.ArchiveStreamFactory;
 import org.apache.commons.compress.archivers.zip.UnsupportedZipFeatureException;
 import org.apache.commons.compress.archivers.zip.ZipArchiveEntry;
 import org.apache.commons.compress.archivers.zip.ZipArchiveInputStream;
@@ -121,7 +120,9 @@ public class DefaultZipContainerDetector implements Detector {
             return TIFF;
         }
         try {
-            String name = ArchiveStreamFactory.detect(new UnsynchronizedByteArrayInputStream(prefix, 0, length));
+            String name =
+                    TikaArchiveStreamFactory.detect(new UnsynchronizedByteArrayInputStream(prefix,
+                            0, length));
             return PackageConstants.getMediaType(name);
         } catch (ArchiveException e) {
             return MediaType.OCTET_STREAM;
diff --git a/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/TikaArchiveStreamFactory.java b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/TikaArchiveStreamFactory.java
new file mode 100644
index 000000000..dc2cb908a
--- /dev/null
+++ b/tika-parsers/tika-parsers-standard/tika-parsers-standard-modules/tika-parser-zip-commons/src/main/java/org/apache/tika/detect/zip/TikaArchiveStreamFactory.java
@@ -0,0 +1,641 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.tika.detect.zip;
+
+import java.io.ByteArrayInputStream;
+import java.io.IOException;
+import java.io.InputStream;
+import java.io.OutputStream;
+import java.security.AccessController;
+import java.security.PrivilegedAction;
+import java.util.Collections;
+import java.util.Locale;
+import java.util.ServiceLoader;
+import java.util.Set;
+import java.util.SortedMap;
+import java.util.TreeMap;
+
+import org.apache.commons.compress.archivers.ArchiveEntry;
+import org.apache.commons.compress.archivers.ArchiveException;
+import org.apache.commons.compress.archivers.ArchiveInputStream;
+import org.apache.commons.compress.archivers.ArchiveOutputStream;
+import org.apache.commons.compress.archivers.ArchiveStreamProvider;
+import org.apache.commons.compress.archivers.StreamingNotSupportedException;
+import org.apache.commons.compress.archivers.ar.ArArchiveInputStream;
+import org.apache.commons.compress.archivers.ar.ArArchiveOutputStream;
+import org.apache.commons.compress.archivers.arj.ArjArchiveInputStream;
+import org.apache.commons.compress.archivers.cpio.CpioArchiveInputStream;
+import org.apache.commons.compress.archivers.cpio.CpioArchiveOutputStream;
+import org.apache.commons.compress.archivers.dump.DumpArchiveInputStream;
+import org.apache.commons.compress.archivers.jar.JarArchiveInputStream;
+import org.apache.commons.compress.archivers.jar.JarArchiveOutputStream;
+import org.apache.commons.compress.archivers.sevenz.SevenZFile;
+import org.apache.commons.compress.archivers.tar.TarArchiveEntry;
+import org.apache.commons.compress.archivers.tar.TarArchiveInputStream;
+import org.apache.commons.compress.archivers.tar.TarArchiveOutputStream;
+import org.apache.commons.compress.archivers.zip.ZipArchiveInputStream;
+import org.apache.commons.compress.archivers.zip.ZipArchiveOutputStream;
+import org.apache.commons.compress.utils.IOUtils;
+import org.apache.commons.compress.utils.Sets;
+
+/**
+ * Factory to create Archive[In|Out]putStreams from names or the first bytes of the InputStream.
+ * In order to add other implementations, you should extend
+ * ArchiveStreamFactory and override the appropriate methods (and call their implementation from
+ * super of course).
+ * <p>
+ * Compressing a ZIP-File:
+ *
+ * <pre>
+ * final OutputStream out = Files.newOutputStream(output.toPath());
+ * ArchiveOutputStream os = new ArchiveStreamFactory().createArchiveOutputStream(ArchiveStreamFactory.ZIP,
+ * out);
+ *
+ * os.putArchiveEntry(new ZipArchiveEntry("testdata/test1.xml"));
+ * IOUtils.copy(Files.newInputStream(file1.toPath()), os);
+ * os.closeArchiveEntry();
+ *
+ * os.putArchiveEntry(new ZipArchiveEntry("testdata/test2.xml"));
+ * IOUtils.copy(Files.newInputStream(file2.toPath()), os);
+ * os.closeArchiveEntry();
+ * os.close();
+ * </pre>
+ * <p>
+ * Decompressing a ZIP-File:
+ *
+ * <pre>
+ * final InputStream is = Files.newInputStream(input.toPath());
+ * ArchiveInputStream in = new ArchiveStreamFactory().createArchiveInputStream(ArchiveStreamFactory.ZIP,
+ * is);
+ * ZipArchiveEntry entry = (ZipArchiveEntry) in.getNextEntry();
+ * OutputStream out = Files.newOutputStream(dir.toPath().resolve(entry.getName()));
+ * IOUtils.copy(in, out);
+ * out.close();
+ * in.close();
+ * </pre>
+ *
+ * @Immutable provided that the deprecated method setEntryEncoding is not used.
+ * @ThreadSafe even if the deprecated method setEntryEncoding is used
+ */
+public class TikaArchiveStreamFactory implements ArchiveStreamProvider {
+
+    private static final int TAR_HEADER_SIZE = 512;
+
+    private static final int TAR_TEST_ENTRY_COUNT = 10;
+
+    private static final int DUMP_SIGNATURE_SIZE = 32;
+
+    private static final int SIGNATURE_SIZE = 12;
+
+    /**
+     * The singleton instance using the platform default encoding.
+     *
+     * @since 1.21
+     */
+    public static final TikaArchiveStreamFactory DEFAULT = new TikaArchiveStreamFactory();
+
+    /**
+     * Constant (value {@value}) used to identify the APK archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String APK = "apk";
+
+    /**
+     * Constant (value {@value}) used to identify the XAPK archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String XAPK = "xapk";
+
+    /**
+     * Constant (value {@value}) used to identify the APKS archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String APKS = "apks";
+
+    /**
+     * Constant (value {@value}) used to identify the APKM archive format.
+     * <p>
+     * APK file extensions are .apk, .xapk, .apks, .apkm
+     * </p>
+     *
+     * @since 1.22
+     */
+    public static final String APKM = "apkm";
+
+    /**
+     * Constant (value {@value}) used to identify the AR archive format.
+     *
+     * @since 1.1
+     */
+    public static final String AR = "ar";
+
+    /**
+     * Constant (value {@value}) used to identify the ARJ archive format. Not supported as an
+     * output stream type.
+     *
+     * @since 1.6
+     */
+    public static final String ARJ = "arj";
+
+    /**
+     * Constant (value {@value}) used to identify the CPIO archive format.
+     *
+     * @since 1.1
+     */
+    public static final String CPIO = "cpio";
+
+    /**
+     * Constant (value {@value}) used to identify the Unix DUMP archive format. Not supported as
+     * an output stream type.
+     *
+     * @since 1.3
+     */
+    public static final String DUMP = "dump";
+
+    /**
+     * Constant (value {@value}) used to identify the JAR archive format.
+     *
+     * @since 1.1
+     */
+    public static final String JAR = "jar";
+
+    /**
+     * Constant used to identify the TAR archive format.
+     *
+     * @since 1.1
+     */
+    public static final String TAR = "tar";
+
+    /**
+     * Constant (value {@value}) used to identify the ZIP archive format.
+     *
+     * @since 1.1
+     */
+    public static final String ZIP = "zip";
+
+    /**
+     * Constant (value {@value}) used to identify the 7z archive format.
+     *
+     * @since 1.8
+     */
+    public static final String SEVEN_Z = "7z";
+
+    private static Iterable<ArchiveStreamProvider> archiveStreamProviderIterable() {
+        return ServiceLoader.load(ArchiveStreamProvider.class, ClassLoader.getSystemClassLoader());
+    }
+
+    /**
+     * Try to determine the type of Archiver
+     *
+     * @param in input stream
+     * @return type of archiver if found
+     * @throws ArchiveException if an archiver cannot be detected in the stream
+     * @since 1.14
+     */
+    public static String detect(final InputStream in) throws ArchiveException {
+        if (in == null) {
+            throw new IllegalArgumentException("Stream must not be null.");
+        }
+
+        if (!in.markSupported()) {
+            throw new IllegalArgumentException("Mark is not supported.");
+        }
+
+        final byte[] signature = new byte[SIGNATURE_SIZE];
+        in.mark(signature.length);
+        int signatureLength = -1;
+        try {
+            signatureLength = IOUtils.readFully(in, signature);
+            in.reset();
+        } catch (final IOException e) {
+            throw new ArchiveException("IOException while reading signature.", e);
+        }
+
+        // For now JAR files are detected as ZIP files.
+        if (ZipArchiveInputStream.matches(signature, signatureLength)) {
+            return ZIP;
+        }
+        // For now JAR files are detected as ZIP files.
+        if (JarArchiveInputStream.matches(signature, signatureLength)) {
+            return JAR;
+        }
+        if (ArArchiveInputStream.matches(signature, signatureLength)) {
+            return AR;
+        }
+        if (CpioArchiveInputStream.matches(signature, signatureLength)) {
+            return CPIO;
+        }
+        if (ArjArchiveInputStream.matches(signature, signatureLength)) {
+            return ARJ;
+        }
+        if (SevenZFile.matches(signature, signatureLength)) {
+            return SEVEN_Z;
+        }
+
+        // Dump needs a bigger buffer to check the signature;
+        final byte[] dumpsig = new byte[DUMP_SIGNATURE_SIZE];
+        in.mark(dumpsig.length);
+        try {
+            signatureLength = IOUtils.readFully(in, dumpsig);
+            in.reset();
+        } catch (final IOException e) {
+            throw new ArchiveException("IOException while reading dump signature", e);
+        }
+        if (DumpArchiveInputStream.matches(dumpsig, signatureLength)) {
+            return DUMP;
+        }
+
+        // Tar needs an even bigger buffer to check the signature; read the first block
+        final byte[] tarHeader = new byte[TAR_HEADER_SIZE];
+        in.mark(tarHeader.length);
+        try {
+            signatureLength = IOUtils.readFully(in, tarHeader);
+            in.reset();
+        } catch (final IOException e) {
+            throw new ArchiveException("IOException while reading tar signature", e);
+        }
+        if (TarArchiveInputStream.matches(tarHeader, signatureLength)) {
+            return TAR;
+        }
+
+        // COMPRESS-117
+        if (signatureLength >= TAR_HEADER_SIZE) {
+            try (TarArchiveInputStream inputStream = new TarArchiveInputStream(
+                    new ByteArrayInputStream(tarHeader))) {
+                // COMPRESS-191 - verify the header checksum
+                // COMPRESS-644 - do not allow zero byte file entries
+                TarArchiveEntry entry = inputStream.getNextEntry();
+                // try to find the first non-directory entry within the first 10 entries.
+                int count = 0;
+                while (entry != null && entry.isDirectory() && entry.isCheckSumOK() &&
+                        count++ < TAR_TEST_ENTRY_COUNT) {
+                    entry = inputStream.getNextEntry();
+                }
+                if (entry != null && entry.isCheckSumOK() && !entry.isDirectory() &&
+                        entry.getSize() > 0 || count > 0) {
+                    return TAR;
+                }
+            } catch (final Exception e) { // NOPMD NOSONAR
+                // can generate IllegalArgumentException as well as IOException auto-detection,
+                // simply not a TAR ignored
+            }
+        }
+        throw new ArchiveException("No Archiver found for the stream signature");
+    }
+
+    /**
+     * Constructs a new sorted map from input stream provider names to provider objects.
+     *
+     * <p>
+     * The map returned by this method will have one entry for each provider for which support is
+     * available in the current Java virtual machine. If two or more
+     * supported provider have the same name then the resulting map will contain just one of them;
+     * which one it will contain is not specified.
+     * </p>
+     *
+     * <p>
+     * The invocation of this method, and the subsequent use of the resulting map, may cause
+     * time-consuming disk or network I/O operations to occur. This method
+     * is provided for applications that need to enumerate all of the available providers, for
+     * example to allow user provider selection.
+     * </p>
+     *
+     * <p>
+     * This method may return different results at different times if new providers are dynamically
+     * made available to the current Java virtual machine.
+     * </p>
+     *
+     * @return An immutable, map from names to provider objects
+     * @since 1.13
+     */
+    public static SortedMap<String, ArchiveStreamProvider> findAvailableArchiveInputStreamProviders() {
+        return AccessController.doPrivileged(
+                (PrivilegedAction<SortedMap<String, ArchiveStreamProvider>>) () -> {
+                    final TreeMap<String, ArchiveStreamProvider> map = new TreeMap<>();
+                    putAll(DEFAULT.getInputStreamArchiveNames(), DEFAULT, map);
+                    archiveStreamProviderIterable().forEach(
+                            provider -> putAll(provider.getInputStreamArchiveNames(), provider,
+                                    map));
+                    return map;
+                });
+    }
+
+    /**
+     * Constructs a new sorted map from output stream provider names to provider objects.
+     *
+     * <p>
+     * The map returned by this method will have one entry for each provider for which support is
+     * available in the current Java virtual machine. If two or more
+     * supported provider have the same name then the resulting map will contain just one of them;
+     * which one it will contain is not specified.
+     * </p>
+     *
+     * <p>
+     * The invocation of this method, and the subsequent use of the resulting map, may cause
+     * time-consuming disk or network I/O operations to occur. This method
+     * is provided for applications that need to enumerate all of the available providers, for
+     * example to allow user provider selection.
+     * </p>
+     *
+     * <p>
+     * This method may return different results at different times if new providers are dynamically
+     * made available to the current Java virtual machine.
+     * </p>
+     *
+     * @return An immutable, map from names to provider objects
+     * @since 1.13
+     */
+    public static SortedMap<String, ArchiveStreamProvider> findAvailableArchiveOutputStreamProviders() {
+        return AccessController.doPrivileged(
+                (PrivilegedAction<SortedMap<String, ArchiveStreamProvider>>) () -> {
+                    final TreeMap<String, ArchiveStreamProvider> map = new TreeMap<>();
+                    putAll(DEFAULT.getOutputStreamArchiveNames(), DEFAULT, map);
+                    archiveStreamProviderIterable().forEach(
+                            provider -> putAll(provider.getOutputStreamArchiveNames(), provider,
+                                    map));
+                    return map;
+                });
+    }
+
+    static void putAll(final Set<String> names, final ArchiveStreamProvider provider,
+                       final TreeMap<String, ArchiveStreamProvider> map) {
+        names.forEach(name -> map.put(toKey(name), provider));
+    }
+
+    private static String toKey(final String name) {
+        return name.toUpperCase(Locale.ROOT);
+    }
+
+    /**
+     * Entry encoding, null for the default.
+     */
+    private volatile String entryEncoding;
+
+    private SortedMap<String, ArchiveStreamProvider> archiveInputStreamProviders;
+
+    private SortedMap<String, ArchiveStreamProvider> archiveOutputStreamProviders;
+
+    /**
+     * Constructs an instance using the platform default encoding.
+     */
+    public TikaArchiveStreamFactory() {
+        this(null);
+    }
+
+    /**
+     * Constructs an instance using the specified encoding.
+     *
+     * @param encoding the encoding to be used.
+     * @since 1.10
+     */
+    public TikaArchiveStreamFactory(final String encoding) {
+        this.entryEncoding = encoding;
+    }
+
+    /**
+     * Creates an archive input stream from an input stream, autodetecting the archive type from the first
+     * few bytes of the stream. The InputStream must support
+     * marks, like BufferedInputStream.
+     *
+     * @param <I> The {@link ArchiveInputStream} type.
+     * @param in  the input stream
+     * @return the archive input stream
+     * @throws ArchiveException               if the archiver name is not known
+     * @throws StreamingNotSupportedException if the format cannot be read from a stream
+     * @throws IllegalArgumentException       if the stream is null or does not support mark
+     */
+    public <I extends ArchiveInputStream<? extends ArchiveEntry>> I createArchiveInputStream(
+            final InputStream in) throws ArchiveException {
+        return createArchiveInputStream(detect(in), in);
+    }
+
+    /**
+     * Creates an archive input stream from an archiver name and an input stream.
+     *
+     * @param <I>          The {@link ArchiveInputStream} type.
+     * @param archiverName the archive name, i.e. {@value #AR}, {@value #ARJ}, {@value #ZIP}, {@value #TAR},
+     * {@value #JAR}, {@value #CPIO}, {@value #DUMP} or
+     *                     {@value #SEVEN_Z}
+     * @param in           the input stream
+     * @return the archive input stream
+     * @throws ArchiveException               if the archiver name is not known
+     * @throws StreamingNotSupportedException if the format cannot be read from a stream
+     * @throws IllegalArgumentException       if the archiver name or stream is null
+     */
+    public <I extends ArchiveInputStream<? extends ArchiveEntry>> I createArchiveInputStream(
+            final String archiverName, final InputStream in) throws ArchiveException {
+        return createArchiveInputStream(archiverName, in, entryEncoding);
+    }
+
+    @SuppressWarnings("unchecked")
+    @Override
+    public <I extends ArchiveInputStream<? extends ArchiveEntry>> I createArchiveInputStream(
+            final String archiverName, final InputStream in, final String actualEncoding)
+            throws ArchiveException {
+
+        if (archiverName == null) {
+            throw new IllegalArgumentException("Archiver name must not be null.");
+        }
+
+        if (in == null) {
+            throw new IllegalArgumentException("InputStream must not be null.");
+        }
+
+        if (AR.equalsIgnoreCase(archiverName)) {
+            return (I) new ArArchiveInputStream(in);
+        }
+        if (ARJ.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new ArjArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new ArjArchiveInputStream(in);
+        }
+        if (ZIP.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new ZipArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new ZipArchiveInputStream(in);
+        }
+        if (TAR.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new TarArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new TarArchiveInputStream(in);
+        }
+        if (JAR.equalsIgnoreCase(archiverName) || APK.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new JarArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new JarArchiveInputStream(in);
+        }
+        if (CPIO.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new CpioArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new CpioArchiveInputStream(in);
+        }
+        if (DUMP.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (I) new DumpArchiveInputStream(in, actualEncoding);
+            }
+            return (I) new DumpArchiveInputStream(in);
+        }
+        if (SEVEN_Z.equalsIgnoreCase(archiverName)) {
+            throw new StreamingNotSupportedException(SEVEN_Z);
+        }
+
+        final ArchiveStreamProvider archiveStreamProvider =
+                getArchiveInputStreamProviders().get(toKey(archiverName));
+        if (archiveStreamProvider != null) {
+            return archiveStreamProvider.createArchiveInputStream(archiverName, in, actualEncoding);
+        }
+
+        throw new ArchiveException("Archiver: " + archiverName + " not found.");
+    }
+
+    /**
+     * Creates an archive output stream from an archiver name and an output stream.
+     *
+     * @param <O>          The {@link ArchiveOutputStream} type.
+     * @param archiverName the archive name, i.e. {@value #AR}, {@value #ZIP}, {@value #TAR},
+     * {@value #JAR} or {@value #CPIO}
+     * @param out          the output stream
+     * @return the archive output stream
+     * @throws ArchiveException               if the archiver name is not known
+     * @throws StreamingNotSupportedException if the format cannot be written to a stream
+     * @throws IllegalArgumentException       if the archiver name or stream is null
+     */
+    public <O extends ArchiveOutputStream<? extends ArchiveEntry>> O createArchiveOutputStream(
+            final String archiverName, final OutputStream out) throws ArchiveException {
+        return createArchiveOutputStream(archiverName, out, entryEncoding);
+    }
+
+    @SuppressWarnings("unchecked")
+    @Override
+    public <O extends ArchiveOutputStream<? extends ArchiveEntry>> O createArchiveOutputStream(
+            final String archiverName, final OutputStream out, final String actualEncoding)
+            throws ArchiveException {
+        if (archiverName == null) {
+            throw new IllegalArgumentException("Archiver name must not be null.");
+        }
+        if (out == null) {
+            throw new IllegalArgumentException("OutputStream must not be null.");
+        }
+
+        if (AR.equalsIgnoreCase(archiverName)) {
+            return (O) new ArArchiveOutputStream(out);
+        }
+        if (ZIP.equalsIgnoreCase(archiverName)) {
+            final ZipArchiveOutputStream zip = new ZipArchiveOutputStream(out);
+            if (actualEncoding != null) {
+                zip.setEncoding(actualEncoding);
+            }
+            return (O) zip;
+        }
+        if (TAR.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (O) new TarArchiveOutputStream(out, actualEncoding);
+            }
+            return (O) new TarArchiveOutputStream(out);
+        }
+        if (JAR.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (O) new JarArchiveOutputStream(out, actualEncoding);
+            }
+            return (O) new JarArchiveOutputStream(out);
+        }
+        if (CPIO.equalsIgnoreCase(archiverName)) {
+            if (actualEncoding != null) {
+                return (O) new CpioArchiveOutputStream(out, actualEncoding);
+            }
+            return (O) new CpioArchiveOutputStream(out);
+        }
+        if (SEVEN_Z.equalsIgnoreCase(archiverName)) {
+            throw new StreamingNotSupportedException(SEVEN_Z);
+        }
+
+        final ArchiveStreamProvider archiveStreamProvider =
+                getArchiveOutputStreamProviders().get(toKey(archiverName));
+        if (archiveStreamProvider != null) {
+            return archiveStreamProvider.createArchiveOutputStream(archiverName, out,
+                    actualEncoding);
+        }
+
+        throw new ArchiveException("Archiver: " + archiverName + " not found.");
+    }
+
+    public SortedMap<String, ArchiveStreamProvider> getArchiveInputStreamProviders() {
+        if (archiveInputStreamProviders == null) {
+            archiveInputStreamProviders =
+                    Collections.unmodifiableSortedMap(findAvailableArchiveInputStreamProviders());
+        }
+        return archiveInputStreamProviders;
+    }
+
+    public SortedMap<String, ArchiveStreamProvider> getArchiveOutputStreamProviders() {
+        if (archiveOutputStreamProviders == null) {
+            archiveOutputStreamProviders =
+                    Collections.unmodifiableSortedMap(findAvailableArchiveOutputStreamProviders());
+        }
+        return archiveOutputStreamProviders;
+    }
+
+    /**
+     * Gets the encoding to use for arj, jar, ZIP, dump, cpio and tar files, or null for the archiver default.
+     *
+     * @return entry encoding, or null for the archiver default
+     * @since 1.5
+     */
+    public String getEntryEncoding() {
+        return entryEncoding;
+    }
+
+    @Override
+    public Set<String> getInputStreamArchiveNames() {
+        return Sets.newHashSet(AR, ARJ, ZIP, TAR, JAR, CPIO, DUMP, SEVEN_Z);
+    }
+
+    @Override
+    public Set<String> getOutputStreamArchiveNames() {
+        return Sets.newHashSet(AR, ZIP, TAR, JAR, CPIO, SEVEN_Z);
+    }
+
+    /**
+     * Sets the encoding to use for arj, jar, ZIP, dump, cpio and tar files. Use null for the archiver default.
+     *
+     * @param entryEncoding the entry encoding, null uses the archiver default.
+     * @since 1.5
+     * @deprecated 1.10 use {@link #ArchiveStreamFactory(String)} to specify the encoding
+     */
+    @Deprecated
+    public void setEntryEncoding(final String entryEncoding) {
+        this.entryEncoding = entryEncoding;
+    }
+
+}