You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hossein Falaki (JIRA)" <ji...@apache.org> on 2016/10/13 18:58:20 UTC
[jira] [Created] (SPARK-17919) Make timeout to RBackend
configurable in SparkR
Hossein Falaki created SPARK-17919:
--------------------------------------
Summary: Make timeout to RBackend configurable in SparkR
Key: SPARK-17919
URL: https://issues.apache.org/jira/browse/SPARK-17919
Project: Spark
Issue Type: Story
Components: SparkR
Affects Versions: 2.0.1
Reporter: Hossein Falaki
I am working on a project where {{gapply()}} is being used with a large dataset that happens to be extremely skewed. On that skewed partition, the user function takes more than 2 hours to return and that turns out to be larger than the timeout that we hardcode in SparkR for backend connection.
{code}
connectBackend <- function(hostname, port, timeout = 6000)
{code}
Ideally user should be able to reconfigure Spark and increase the timeout. It should be a small fix.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org