You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "Hive QA (JIRA)" <ji...@apache.org> on 2016/11/10 03:18:58 UTC

[jira] [Commented] (HIVE-14453) refactor physical writing of ORC data and metadata to FS from the logical writers

    [ https://issues.apache.org/jira/browse/HIVE-14453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15652874#comment-15652874 ] 

Hive QA commented on HIVE-14453:
--------------------------------



Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12838280/HIVE-14453.02.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 10637 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid] (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats] (batchId=145)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2058/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2058/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2058/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12838280 - PreCommit-HIVE-Build

> refactor physical writing of ORC data and metadata to FS from the logical writers
> ---------------------------------------------------------------------------------
>
>                 Key: HIVE-14453
>                 URL: https://issues.apache.org/jira/browse/HIVE-14453
>             Project: Hive
>          Issue Type: Bug
>            Reporter: Sergey Shelukhin
>            Assignee: Sergey Shelukhin
>         Attachments: HIVE-14453.01.patch, HIVE-14453.02.patch, HIVE-14453.patch
>
>
> ORC data doesn't have to go directly into an HDFS stream via buffers, it can go somewhere else (e.g. a write-thru cache, or an addressable system that doesn't require the stream blocks to be held in memory before writing them all together).
> To that effect, it would be nice to abstract the data block/metadata structure creating from the physical file concerns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)