You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Tomasz Bartczak (JIRA)" <ji...@apache.org> on 2019/01/15 20:24:00 UTC
[jira] [Created] (SPARK-26627) sql_ctx loses '_conf' attribute for
a pyspark dataframe converted to jdf and back
Tomasz Bartczak created SPARK-26627:
---------------------------------------
Summary: sql_ctx loses '_conf' attribute for a pyspark dataframe converted to jdf and back
Key: SPARK-26627
URL: https://issues.apache.org/jira/browse/SPARK-26627
Project: Spark
Issue Type: Bug
Components: PySpark
Affects Versions: 2.4.0
Reporter: Tomasz Bartczak
Having a pyspark code:
{code:java}
raw = spark.read.parquet("")
converted = DataFrame(raw._jdf, spark)
converted.toPandas(){code}
what I get when running toPandas is:
{code:java}
-> 2079 if self.sql_ctx._conf.pandasRespectSessionTimeZone():
2080 timezone = self.sql_ctx._conf.sessionLocalTimeZone()
2081 else:
AttributeError: 'SparkSession' object has no attribute '_conf'{code}
So it looks like after converting df to java version and back sql_ctx lost '_conf' attribute.
Of course in reality life this raw._jdf is passed to a specific jvm method to do some operations on it.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org