You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@carbondata.apache.org by "yangxy (JIRA)" <ji...@apache.org> on 2017/05/05 06:43:04 UTC

[jira] [Updated] (CARBONDATA-1025) dependency univocity-parsers conflict with spark2.1

     [ https://issues.apache.org/jira/browse/CARBONDATA-1025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

yangxy updated CARBONDATA-1025:
-------------------------------
    Summary: dependency univocity-parsers conflict with spark2.1  (was: dependency univocity-parsers conflict whith spark2.1)

> dependency univocity-parsers conflict with spark2.1
> ---------------------------------------------------
>
>                 Key: CARBONDATA-1025
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1025
>             Project: CarbonData
>          Issue Type: Bug
>            Reporter: yangxy
>
> Caused by: java.lang.NoSuchMethodError: com.univocity.parsers.csv.CsvParserSettings.setUnescapedQuoteHandling(Lcom/univocity/parsers/csv/UnescapedQuoteHandling;)V
>   at org.apache.spark.sql.execution.datasources.csv.CsvReader.<init>(CSVParser.scala:47)
>   at org.apache.spark.sql.execution.datasources.csv.CSVFileFormat$$anonfun$buildReader$1.apply(CSVFileFormat.scala:158)
>   at org.apache.spark.sql.execution.datasources.csv.CSVFileFormat$$anonfun$buildReader$1.apply(CSVFileFormat.scala:146)
>   at org.apache.spark.sql.execution.datasources.FileFormat$$anon$1.apply(FileFormat.scala:138)
>   at org.apache.spark.sql.execution.datasources.FileFormat$$anon$1.apply(FileFormat.scala:122)
>   at org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.nextIterator(FileScanRDD.scala:168)
>   at org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:109)
>   at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithoutKey$(Unknown Source)
>   at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
>   at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322)
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)