You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@kylin.apache.org by "Shaofeng SHI (JIRA)" <ji...@apache.org> on 2017/09/25 10:15:00 UTC

[jira] [Comment Edited] (KYLIN-2788) HFile is not written to S3

    [ https://issues.apache.org/jira/browse/KYLIN-2788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16178809#comment-16178809 ] 

Shaofeng SHI edited comment on KYLIN-2788 at 9/25/17 10:14 AM:
---------------------------------------------------------------

Attache the configuration parameters on this step. We can see that it uses "org.apache.hadoop.mapred.DirectFileOutputCommitter" as the committer class.

But in HBase's HFileOutputFormat2.java is using old mapreduce API. Which may need use "
org.apache.hadoop.mapreduce.lib.output.DirectFileOutputCommitter" as the committer

On EMR, we can see there are two types of the direct output committer:
/usr/lib/hbase/lib/hadoop-mapreduce-client-core.jar - org/apache/hadoop/mapred/DirectFileOutputCommitter.class
/usr/lib/hbase/lib/hadoop-mapreduce-client-core.jar - org/apache/hadoop/mapreduce/lib/output/DirectFileOutputCommitter.class
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-core-2.7.3-amzn-2.jar - org/apache/hadoop/mapred/DirectFileOutputCommitter.class
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-core-2.7.3-amzn-2.jar - org/apache/hadoop/mapreduce/lib/output/DirectFileOutputCommitter.class


was (Author: shaofengshi):
Attache the configuration parameters on this step. We can see that it uses "org.apache.hadoop.mapred.DirectFileOutputCommitter" as the committer class.

But in HBase's HFileOutputFormat2.java is using old mapreduce API. Which may need use "
org.apache.hadoop.mapreduce.lib.output.DirectFileOutputCommitter" as the committer

> HFile is not written to S3
> --------------------------
>
>                 Key: KYLIN-2788
>                 URL: https://issues.apache.org/jira/browse/KYLIN-2788
>             Project: Kylin
>          Issue Type: Bug
>    Affects Versions: v2.0.0
>            Reporter: Alexander Sterligov
>         Attachments: ConvertoHFileOnS3.xlsx
>
>
> I set kylin.hbase.cluster.fs to s3 bucket where hbase lives.
> Step "Convert Cuboid Data to HFile" finished without errors. Statistics at the end of the job said that it has written lot's of data to s3.
> But there is no hfiles in kylin_metadata folder (kylin_metadata /kylin-1e436685-7102-4621-a4cb-6472b866126d/<table name>/hfile), but only _temporary folder and _SUCCESS file.
> _temporary contains hfiles inside attempt folders. it looks like there were not copied from _temporary to result dir. But there is no errors neither in kylin log, nor in reducers' logs.
> Then loading empty hfiles produces empty segments.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)