You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hossein Falaki (JIRA)" <ji...@apache.org> on 2016/10/13 18:58:20 UTC

[jira] [Created] (SPARK-17919) Make timeout to RBackend configurable in SparkR

Hossein Falaki created SPARK-17919:
--------------------------------------

             Summary: Make timeout to RBackend configurable in SparkR
                 Key: SPARK-17919
                 URL: https://issues.apache.org/jira/browse/SPARK-17919
             Project: Spark
          Issue Type: Story
          Components: SparkR
    Affects Versions: 2.0.1
            Reporter: Hossein Falaki


I am working on a project where {{gapply()}} is being used with a large dataset that happens to be extremely skewed. On that skewed partition, the user function takes more than 2 hours to return and that turns out to be larger than the timeout that we hardcode in SparkR for backend connection.

{code}
connectBackend <- function(hostname, port, timeout = 6000) 
{code}

Ideally user should be able to reconfigure Spark and increase the timeout. It should be a small fix.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org