You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Nikhil Goyal <no...@gmail.com> on 2018/10/22 18:48:29 UTC

Writing to vertica from spark

Hi guys,

My code is failing with this error

java.lang.Exception: S2V: FATAL ERROR for job S2V_job9197956021769393773.
Job status information is available in the Vertica table
test.S2V_JOB_STATUS_USER_NGOYAL.  Unable to save intermediate orc files to
HDFS path:hdfs://hadoop-dw2-nn.smf1.com/tmp/S2V_job9197956021769393773.
*Error message:org.apache.spark.sql.AnalysisException: The ORC data source
must be used with Hive support enabled;*

This is how I am writing it. I followed steps from this link
<https://www.vertica.com/blog/integrating-apache-spark/>.

dataFrame
  .write
  .format("com.vertica.spark.datasource.DefaultSource")
  .options(connectionProperties)
  .mode(SaveMode.Append)
  .save()

Does anybody have any idea how to fix this?


Thanks
Nikhil

Re: Writing to vertica from spark

Posted by Nikhil Goyal <no...@gmail.com>.
Fixed this by setting fileformat -> "parquet"

On Mon, Oct 22, 2018 at 11:48 AM Nikhil Goyal <no...@gmail.com> wrote:

> Hi guys,
>
> My code is failing with this error
>
> java.lang.Exception: S2V: FATAL ERROR for job S2V_job9197956021769393773.
> Job status information is available in the Vertica table
> test.S2V_JOB_STATUS_USER_NGOYAL.  Unable to save intermediate orc files to
> HDFS path:hdfs://hadoop-dw2-nn.smf1.com/tmp/S2V_job9197956021769393773.
> *Error message:org.apache.spark.sql.AnalysisException: The ORC data source
> must be used with Hive support enabled;*
>
> This is how I am writing it. I followed steps from this link
> <https://www.vertica.com/blog/integrating-apache-spark/>.
>
> dataFrame
>   .write
>   .format("com.vertica.spark.datasource.DefaultSource")
>   .options(connectionProperties)
>   .mode(SaveMode.Append)
>   .save()
>
> Does anybody have any idea how to fix this?
>
>
> Thanks
> Nikhil
>
>