You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sen Fang (JIRA)" <ji...@apache.org> on 2015/10/21 21:42:27 UTC

[jira] [Created] (SPARK-11244) sparkR.stop doesn't clean up .sparkRSQLsc in environment

Sen Fang created SPARK-11244:
--------------------------------

             Summary: sparkR.stop doesn't clean up .sparkRSQLsc in environment
                 Key: SPARK-11244
                 URL: https://issues.apache.org/jira/browse/SPARK-11244
             Project: Spark
          Issue Type: Bug
          Components: SparkR
    Affects Versions: 1.5.1
            Reporter: Sen Fang


Currently {{sparkR.stop}} removes relevant variables from {{.sparkREnv}} for SparkContext and backend. However it doesn't clean up {{.sparkRSQLsc}} and {{.sparkRHivesc}}.

It results 
{code}
sc <- sparkR.init("local")
sqlContext <- sparkRSQL.init(sc)
sparkR.stop()
sc <- sparkR.init("local")
sqlContext <- sparkRSQL.init(sc)
sqlContext
{code}
producing
{code}
 sqlContext
Error in callJMethod(x, "getClass") : 
  Invalid jobj 1. If SparkR was restarted, Spark operations need to be re-executed.
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org