You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sergei Lebedev (JIRA)" <ji...@apache.org> on 2017/01/24 21:10:26 UTC

[jira] [Created] (SPARK-19353) Support binary I/O in PipedRDD

Sergei Lebedev created SPARK-19353:
--------------------------------------

             Summary: Support binary I/O in PipedRDD
                 Key: SPARK-19353
                 URL: https://issues.apache.org/jira/browse/SPARK-19353
             Project: Spark
          Issue Type: Improvement
            Reporter: Sergei Lebedev
            Priority: Minor


The current design of RDD.pipe is very restrictive. 

It is line-based, each element of the input RDD [gets serialized|https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/rdd/PipedRDD.scala#L143] into one or more lines. Similarly for the output of the child process, one line corresponds to a single element of the output RDD. 

It allows to customize the output format via {{printRDDElement}}, but not the input format.

It is not designed for extensibility. The only way to get a "BinaryPipedRDD" is to copy/paste most of it and change the relevant parts.

These limitations have been discussed on [SO|http://stackoverflow.com/questions/27986830/how-to-pipe-binary-data-in-apache-spark] and the mailing list, but alas no issue has been created.

A possible solution to at least the first two limitations is to factor out the format into a separate object (or objects). For instance, {{InputWriter}} and {{OutputReader}}, following Hadoop streaming API. 

{code}
trait InputWriter[T] {
    def write(os: OutputStream, elem: T)
}

trait OutputReader[T] {
    def read(is: InputStream): T
}
{code}

The default configuration would be to write and read in line-based format, but the users will also be able to selectively swap those to the appropriate implementations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org