You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dongjoon Hyun (Jira)" <ji...@apache.org> on 2021/03/18 22:26:00 UTC
[jira] [Resolved] (SPARK-34792) Restore previous behaviour of
randomSplit from spark-2.4.7 in spark-3
[ https://issues.apache.org/jira/browse/SPARK-34792?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Dongjoon Hyun resolved SPARK-34792.
-----------------------------------
Resolution: Not A Problem
Hi, [~kondziolka9ld].
This doesn't look like a bug. It can be different in many reasons. Not only Spark's code, but also there is difference between Scala 2.11 and Scala 2.12. BTW, please send your question to dev@spark.apache.org next time.
> Restore previous behaviour of randomSplit from spark-2.4.7 in spark-3
> ---------------------------------------------------------------------
>
> Key: SPARK-34792
> URL: https://issues.apache.org/jira/browse/SPARK-34792
> Project: Spark
> Issue Type: Question
> Components: Spark Core, SQL
> Affects Versions: 3.0.1
> Reporter: kondziolka9ld
> Priority: Major
>
> Hi,
> Please consider a following difference of `randomSplit` method even despite of the same seed.
>
> {code:java}
> ____ __
> / __/__ ___ _____/ /__
> _\ \/ _ \/ _ `/ __/ '_/
> /___/ .__/\_,_/_/ /_/\_\ version 2.4.7
> /_/
>
> Using Scala version 2.12.10 (OpenJDK 64-Bit Server VM, Java 1.8.0_282)
> Type in expressions to have them evaluated.
> Type :help for more information.
> scala> val Array(f, s) = Seq(1,2,3,4,5,6,7,8,9,10).toDF.randomSplit(Array(0.3, 0.7), 42)
> f: org.apache.spark.sql.Dataset[org.apache.spark.sql.Row] = [value: int]
> s: org.apache.spark.sql.Dataset[org.apache.spark.sql.Row] = [value: int]
> scala> f.show
> +-----+
> |value|
> +-----+
> | 4|
> +-----+
> scala> s.show
> +-----+
> |value|
> +-----+
> | 1|
> | 2|
> | 3|
> | 5|
> | 6|
> | 7|
> | 8|
> | 9|
> | 10|
> +-----+
> {code}
> while as on spark-3
> {code:java}
> scala> val Array(f, s) = Seq(1,2,3,4,5,6,7,8,9,10).toDF.randomSplit(Array(0.3, 0.7), 42)
> f: org.apache.spark.sql.Dataset[org.apache.spark.sql.Row] = [value: int]
> s: org.apache.spark.sql.Dataset[org.apache.spark.sql.Row] = [value: int]
> scala> f.show
> +-----+
> |value|
> +-----+
> | 5|
> | 10|
> +-----+
> scala> s.show
> +-----+
> |value|
> +-----+
> | 1|
> | 2|
> | 3|
> | 4|
> | 6|
> | 7|
> | 8|
> | 9|
> +-----+
> {code}
> I guess that implementation of `sample` method changed.
> Is it possible to restore previous behaviour?
> Thanks in advance!
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org