You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hyukjin Kwon (Jira)" <ji...@apache.org> on 2020/03/11 06:34:00 UTC
[jira] [Resolved] (SPARK-31057) approxQuantile function of spark ,
not taking List as first parameter
[ https://issues.apache.org/jira/browse/SPARK-31057?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Hyukjin Kwon resolved SPARK-31057.
----------------------------------
Resolution: Invalid
> approxQuantile function of spark , not taking List<String> as first parameter
> ------------------------------------------------------------------------------
>
> Key: SPARK-31057
> URL: https://issues.apache.org/jira/browse/SPARK-31057
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 2.4.1, 2.4.3
> Environment: spark-sql-2.4.3 v and eclipse neon ide.
>
> Reporter: Shyam
> Priority: Major
>
> 0
> I am using spark-sql-2.4.1v in my project with java8.
> I need to calculate the quantiles on the some of the (calculated) columns (i.e. con_dist_1 , con_dist_2 ) of below given dataframe df.
> {{List<String> calcColmns = Arrays.asList("con_dist_1","con_dist_2")}}
> When I am trying to use first version of approxQuantile i.e. approxQuantile(List, List, double) as below
> Dataset<Row> df = //dataset
> {{List<List<Double>> quants = df.stat().approxQuantile(calcColmns , Array(0.0,0.1,0.5),0.0);}}
> *It is giving error :*
> {quote}The method approxQuantile(String, double[], double) in the type DataFrameStatFunctions is not applicable for the arguments (List, List, double)
> {quote}
> so what is wrong here , I am doing it in my eclipseIDE. Why it is not invoking List even though i am passing List ??
> Really appreciate any help on this.
> more details are added here
> [https://stackoverflow.com/questions/60550152/issue-with-approxquantile-of-spark-not-recognizing-liststring]
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org