You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hyukjin Kwon (JIRA)" <ji...@apache.org> on 2017/10/08 16:36:00 UTC
[jira] [Resolved] (SPARK-17145) Object with many fields causes Seq
Serialization Bug
[ https://issues.apache.org/jira/browse/SPARK-17145?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Hyukjin Kwon resolved SPARK-17145.
----------------------------------
Resolution: Cannot Reproduce
The test at least looks passing in the master. I am resolving this as a {{Cannot Reproduce}}.
> Object with many fields causes Seq Serialization Bug
> -----------------------------------------------------
>
> Key: SPARK-17145
> URL: https://issues.apache.org/jira/browse/SPARK-17145
> Project: Spark
> Issue Type: Bug
> Affects Versions: 2.0.0
> Environment: OS: OSX El Capitan 10.11.6
> Reporter: Abdulla Al-Qawasmeh
>
> The unit test here (https://gist.github.com/abdulla16/433faf7df59fce11a5fff284bac0d945) describes the problem.
> It looks like Spark is having problems serializing a Scala Seq when it's part of an object with many fields (I'm not 100% sure it's a serialization problem). The deserialized Seq ends up with as many items as the original Seq, however, all the items are copies of the last item in the original Seq.
> The object that I used in my unit test (as an example) is a Tuple5. However, I've seen this behavior in other types of objects.
> Reducing MyClass5 to only two fields (field34 and field35) causes the unit test to pass.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org