You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Tomasz Bartczak (JIRA)" <ji...@apache.org> on 2019/01/15 20:24:00 UTC

[jira] [Created] (SPARK-26627) sql_ctx loses '_conf' attribute for a pyspark dataframe converted to jdf and back

Tomasz Bartczak created SPARK-26627:
---------------------------------------

             Summary: sql_ctx loses '_conf' attribute for a pyspark dataframe converted to jdf and back
                 Key: SPARK-26627
                 URL: https://issues.apache.org/jira/browse/SPARK-26627
             Project: Spark
          Issue Type: Bug
          Components: PySpark
    Affects Versions: 2.4.0
            Reporter: Tomasz Bartczak


Having a pyspark code:
{code:java}
raw = spark.read.parquet("")
converted = DataFrame(raw._jdf, spark)
converted.toPandas(){code}
what I get when running toPandas is:
{code:java}
-> 2079         if self.sql_ctx._conf.pandasRespectSessionTimeZone():
   2080             timezone = self.sql_ctx._conf.sessionLocalTimeZone()
   2081         else:

AttributeError: 'SparkSession' object has no attribute '_conf'{code}
So it looks like after converting df to java version and back sql_ctx lost '_conf' attribute.

 

Of course in reality life this raw._jdf is passed to a specific jvm method to do some operations on it.

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org