You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hyukjin Kwon (JIRA)" <ji...@apache.org> on 2017/10/08 16:36:00 UTC

[jira] [Resolved] (SPARK-17145) Object with many fields causes Seq Serialization Bug

     [ https://issues.apache.org/jira/browse/SPARK-17145?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Hyukjin Kwon resolved SPARK-17145.
----------------------------------
    Resolution: Cannot Reproduce

The test at least looks passing in the master. I am resolving this as a {{Cannot Reproduce}}.

> Object with many fields causes Seq Serialization Bug 
> -----------------------------------------------------
>
>                 Key: SPARK-17145
>                 URL: https://issues.apache.org/jira/browse/SPARK-17145
>             Project: Spark
>          Issue Type: Bug
>    Affects Versions: 2.0.0
>         Environment: OS: OSX El Capitan 10.11.6
>            Reporter: Abdulla Al-Qawasmeh
>
> The unit test here (https://gist.github.com/abdulla16/433faf7df59fce11a5fff284bac0d945) describes the problem. 
> It looks like Spark is having problems serializing a Scala Seq when it's part of an object with many fields (I'm not 100% sure it's a serialization problem). The deserialized Seq ends up with as many items as the original Seq, however, all the items are copies of the last item in the original Seq.
> The object that I used in my unit test (as an example) is a Tuple5. However, I've seen this behavior in other types of objects. 
> Reducing MyClass5 to only two fields (field34 and field35) causes the unit test to pass. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org