You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Xiangrui Meng (JIRA)" <ji...@apache.org> on 2018/12/19 17:37:00 UTC

[jira] [Created] (SPARK-26410) Support per Pandas UDF configuration

Xiangrui Meng created SPARK-26410:
-------------------------------------

             Summary: Support per Pandas UDF configuration
                 Key: SPARK-26410
                 URL: https://issues.apache.org/jira/browse/SPARK-26410
             Project: Spark
          Issue Type: New Feature
          Components: PySpark
    Affects Versions: 3.0.0
            Reporter: Xiangrui Meng


We use a "maxRecordsPerBatch" conf to control the batch sizes. However, the "right" batch size usually depends on the task itself. It would be nice if user can configure the batch size when they declare the Pandas UDF.

This is orthogonal to SPARK-23258 (using max buffer size instead of row count).

Besides API, we should also discuss how to merge Pandas UDFs of different configurations. For example,

{code}
df.select(predict1(col("features"), predict2(col("features")))
{code}

when predict1 requests 100 rows per batch, while predict2 requests 120 rows per batch.

cc: [~icexelloss] [~bryanc] [~holdenk] [~ueshin] [~smilegator]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org