You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Andrew Davidson (JIRA)" <ji...@apache.org> on 2015/12/22 21:11:46 UTC
[jira] [Updated] (SPARK-12483) Data Frame as() does not work in
Java
[ https://issues.apache.org/jira/browse/SPARK-12483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Andrew Davidson updated SPARK-12483:
------------------------------------
Attachment: SPARK_12483_unitTest.java
add a unit test file
> Data Frame as() does not work in Java
> -------------------------------------
>
> Key: SPARK-12483
> URL: https://issues.apache.org/jira/browse/SPARK-12483
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 1.5.2
> Environment: Mac El Cap 10.11.2
> Java 8
> Reporter: Andrew Davidson
> Attachments: SPARK_12483_unitTest.java
>
>
> Following unit test demonstrates a bug in as(). The column name for aliasDF was not changed
> @Test
> public void bugDataFrameAsTest() {
> DataFrame df = createData();
> df.printSchema();
> df.show();
>
> DataFrame aliasDF = df.select("id").as("UUID");
> aliasDF.printSchema();
> aliasDF.show();
> }
> DataFrame createData() {
> Features f1 = new Features(1, category1);
> Features f2 = new Features(2, category2);
> ArrayList<Features> data = new ArrayList<Features>(2);
> data.add(f1);
> data.add(f2);
> //JavaRDD<Features> rdd = javaSparkContext.parallelize(Arrays.asList(f1, f2));
> JavaRDD<Features> rdd = javaSparkContext.parallelize(data);
> DataFrame df = sqlContext.createDataFrame(rdd, Features.class);
> return df;
> }
> This is the output I got (without the spark log msgs)
> root
> |-- id: integer (nullable = false)
> |-- labelStr: string (nullable = true)
> +---+------------+
> | id| labelStr|
> +---+------------+
> | 1| noise|
> | 2|questionable|
> +---+------------+
> root
> |-- id: integer (nullable = false)
> +---+
> | id|
> +---+
> | 1|
> | 2|
> +---+
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org