You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hbase.apache.org by "stack (JIRA)" <ji...@apache.org> on 2018/02/06 01:22:00 UTC

[jira] [Reopened] (HBASE-19841) Tests against hadoop3 fail with StreamLacksCapabilityException

     [ https://issues.apache.org/jira/browse/HBASE-19841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

stack reopened HBASE-19841:
---------------------------

Reopening. Breaks launching of MR jobs on a cluster. Here is what a good launch looks like:
{code}
...
18/02/05 17:11:33 INFO impl.YarnClientImpl: Submitted application application_1517369646236_0009
18/02/05 17:11:33 INFO mapreduce.Job: The url to track the job: http://ve0524.halxg.cloudera.com:10134/proxy/application_1517369646236_0009/
18/02/05 17:11:33 INFO mapreduce.Job: Running job: job_1517369646236_0009
18/02/05 17:11:40 INFO mapreduce.Job: Job job_1517369646236_0009 running in uber mode : false
18/02/05 17:11:40 INFO mapreduce.Job:  map 0% reduce 0%
18/02/05 17:11:57 INFO mapreduce.Job:  map 14% reduce 0%
...
{code}

... but now it does this....

{code}
18/02/05 17:17:54 INFO mapreduce.Job: The url to track the job: http://ve0524.halxg.cloudera.com:10134/proxy/application_1517369646236_0011/
18/02/05 17:17:54 INFO mapreduce.Job: Running job: job_1517369646236_0011
18/02/05 17:17:56 INFO mapreduce.Job: Job job_1517369646236_0011 running in uber mode : false
18/02/05 17:17:56 INFO mapreduce.Job:  map 0% reduce 0%
18/02/05 17:17:56 INFO mapreduce.Job: Job job_1517369646236_0011 failed with state FAILED due to: Application application_1517369646236_0011 failed 2 times due to AM Container for appattempt_1517369646236_0011_000002 exited with  exitCode: -1000
Failing this attempt.Diagnostics: File file:/tmp/stack/.staging/job_1517369646236_0011/job.splitmetainfo does not exist
java.io.FileNotFoundException: File file:/tmp/stack/.staging/job_1517369646236_0011/job.splitmetainfo does not exist
        at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:635)
        at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:861)
        at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:625)
        at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:442)
        at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:253)
        at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:63)
        at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:361)
        at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
        at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:358)
        at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:62)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)

For more detailed output, check the application tracking page: http://ve0524.halxg.cloudera.com:8188/applicationhistory/app/application_1517369646236_0011 Then click on links to logs of each attempt.
. Failing the application.
18/02/05 17:17:56 INFO mapreduce.Job: Counters: 0
{code}

If I revert this patch, the submit runs again.

I'd made the staging dir /tmp/stack and seemed to get further... The job staging was made in the local fs... but it seems like we are then looking for it up in hdfs. My guess is our stamping the fs as local until minihdfscluster starts works for the unit test case but it messes up the inference of fs that allows the above submission to work.

I'd like to revert this if thats ok.

> Tests against hadoop3 fail with StreamLacksCapabilityException
> --------------------------------------------------------------
>
>                 Key: HBASE-19841
>                 URL: https://issues.apache.org/jira/browse/HBASE-19841
>             Project: HBase
>          Issue Type: Test
>            Reporter: Ted Yu
>            Assignee: Mike Drob
>            Priority: Major
>             Fix For: 2.0.0-beta-2
>
>         Attachments: 19841.007.patch, 19841.06.patch, 19841.v0.txt, 19841.v1.txt, HBASE-19841.v10.patch, HBASE-19841.v11.patch, HBASE-19841.v11.patch, HBASE-19841.v2.patch, HBASE-19841.v3.patch, HBASE-19841.v4.patch, HBASE-19841.v5.patch, HBASE-19841.v7.patch, HBASE-19841.v8.patch, HBASE-19841.v8.patch, HBASE-19841.v8.patch, HBASE-19841.v9.patch
>
>
> The following can be observed running against hadoop3:
> {code}
> java.io.IOException: cannot get log writer
> 	at org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.compactingSetUp(TestCompactingMemStore.java:107)
> 	at org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.setUp(TestCompactingMemStore.java:89)
> Caused by: org.apache.hadoop.hbase.util.CommonFSUtils$StreamLacksCapabilityException: hflush and hsync
> 	at org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.compactingSetUp(TestCompactingMemStore.java:107)
> 	at org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.setUp(TestCompactingMemStore.java:89)
> {code}
> This was due to hbase-server/src/test/resources/hbase-site.xml not being picked up by Configuration object. Among the configs from this file, the value for "hbase.unsafe.stream.capability.enforce" relaxes check for presence of hflush and hsync. Without this config entry,  StreamLacksCapabilityException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)