You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@toree.apache.org by "Hollin Wilkins (JIRA)" <ji...@apache.org> on 2016/10/02 17:31:20 UTC

[jira] [Comment Edited] (TOREE-345) Spark 2.0.0 Fails when trying to use spark-avro

    [ https://issues.apache.org/jira/browse/TOREE-345?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15540697#comment-15540697 ] 

Hollin Wilkins edited comment on TOREE-345 at 10/2/16 5:30 PM:
---------------------------------------------------------------

Found out why the error happens, slf4j as a dependency to spark-avro causes it for some reason. When including spark-avro as a spark package via __TOREE_SPARK_OPTS__ in the kernel.json file, the error goes away because Spark Packages will automatically evict the slf4j dependency.

Output from including Spark-Avro via Spark Packages
{code}
:: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
	confs: [default]
	found com.databricks#spark-avro_2.11;3.0.1 in list
	found org.slf4j#slf4j-api;1.7.5 in list
	found org.apache.avro#avro;1.7.6 in list
	found org.codehaus.jackson#jackson-core-asl;1.9.13 in list
	found org.codehaus.jackson#jackson-mapper-asl;1.9.13 in list
	found com.thoughtworks.paranamer#paranamer;2.3 in list
	found org.xerial.snappy#snappy-java;1.0.5 in list
	found org.apache.commons#commons-compress;1.4.1 in list
	found org.tukaani#xz;1.0 in list
:: resolution report :: resolve 341ms :: artifacts dl 5ms
	:: modules in use:
	com.databricks#spark-avro_2.11;3.0.1 from list in [default]
	com.thoughtworks.paranamer#paranamer;2.3 from list in [default]
	org.apache.avro#avro;1.7.6 from list in [default]
	org.apache.commons#commons-compress;1.4.1 from list in [default]
	org.codehaus.jackson#jackson-core-asl;1.9.13 from list in [default]
	org.codehaus.jackson#jackson-mapper-asl;1.9.13 from list in [default]
	org.slf4j#slf4j-api;1.7.5 from list in [default]
	org.tukaani#xz;1.0 from list in [default]
	org.xerial.snappy#snappy-java;1.0.5 from list in [default]
	:: evicted modules:
	org.slf4j#slf4j-api;1.6.4 by [org.slf4j#slf4j-api;1.7.5] in [default]
{code}


was (Author: hollinwilkins):
Found out why the error happens, slf4j as a dependency to spark-avro causes it for some reason. When including spark-avro as a spark package via Spark options in the kernel.json file, the error goes away because Spark Packages will automatically evict the dlf4j dependency.

Output from including Spark-Avro via Spark Packages
{code}
:: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
	confs: [default]
	found com.databricks#spark-avro_2.11;3.0.1 in list
	found org.slf4j#slf4j-api;1.7.5 in list
	found org.apache.avro#avro;1.7.6 in list
	found org.codehaus.jackson#jackson-core-asl;1.9.13 in list
	found org.codehaus.jackson#jackson-mapper-asl;1.9.13 in list
	found com.thoughtworks.paranamer#paranamer;2.3 in list
	found org.xerial.snappy#snappy-java;1.0.5 in list
	found org.apache.commons#commons-compress;1.4.1 in list
	found org.tukaani#xz;1.0 in list
:: resolution report :: resolve 341ms :: artifacts dl 5ms
	:: modules in use:
	com.databricks#spark-avro_2.11;3.0.1 from list in [default]
	com.thoughtworks.paranamer#paranamer;2.3 from list in [default]
	org.apache.avro#avro;1.7.6 from list in [default]
	org.apache.commons#commons-compress;1.4.1 from list in [default]
	org.codehaus.jackson#jackson-core-asl;1.9.13 from list in [default]
	org.codehaus.jackson#jackson-mapper-asl;1.9.13 from list in [default]
	org.slf4j#slf4j-api;1.7.5 from list in [default]
	org.tukaani#xz;1.0 from list in [default]
	org.xerial.snappy#snappy-java;1.0.5 from list in [default]
	:: evicted modules:
	org.slf4j#slf4j-api;1.6.4 by [org.slf4j#slf4j-api;1.7.5] in [default]
{code}

> Spark 2.0.0 Fails when trying to use spark-avro
> -----------------------------------------------
>
>                 Key: TOREE-345
>                 URL: https://issues.apache.org/jira/browse/TOREE-345
>             Project: TOREE
>          Issue Type: Bug
>         Environment: Spark 2.0.0, Scala 2.11, Hadoop 2.7, Toree Git SHA: 7c1bfb6df7130477c558e69bbb518b0af364e06a
>            Reporter: Hollin Wilkins
>
> When trying to use the spark-avro project to load Avro files from Jupyter, we get errors.
> First:
> {code}
> %AddDeps com.databricks spark-avro_2.11 3.0.1 --transitive --trace
> {code}
> Then try to load an Avro file and show it:
> {code}
> spark.sqlContext.read.format("com.databricks.spark.avro").load("/tmp/test.avro").show()
> {code}
> And we get an error. I will attach the trace as a file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)