You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-issues@hadoop.apache.org by "Karthik Kambatla (JIRA)" <ji...@apache.org> on 2014/07/30 02:24:39 UTC

[jira] [Moved] (MAPREDUCE-6012) DBInputSplit creates invalid ranges on Oracle

     [ https://issues.apache.org/jira/browse/MAPREDUCE-6012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Karthik Kambatla moved HADOOP-9530 to MAPREDUCE-6012:
-----------------------------------------------------

    Affects Version/s:     (was: 1.1.2)
                       1.2.1
                  Key: MAPREDUCE-6012  (was: HADOOP-9530)
              Project: Hadoop Map/Reduce  (was: Hadoop Common)

> DBInputSplit creates invalid ranges on Oracle
> ---------------------------------------------
>
>                 Key: MAPREDUCE-6012
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6012
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>    Affects Versions: 1.2.1
>            Reporter: Julien Serdaru
>            Assignee: Wei Yan
>         Attachments: HADOOP-9530.patch
>
>
> The DBInputFormat on Oracle does not create valid ranges.
> The method getSplit line 263 is as follows:
>           split = new DBInputSplit(i * chunkSize, (i * chunkSize) + chunkSize);
> So the first split will have a start value of 0 (0*chunkSize).
> However, the OracleDBRecordReader, line 84 is as follows:
>       if (split.getLength() > 0 && split.getStart() > 0){
> Since the start value of the first range is equal to 0, we will skip the block that partitions the input set. As a result, one of the map task will process the entire data set, rather than the partition.
> I'm assuming the fix is trivial and would involve removing the second check in the if block.
> Also, I believe the OracleDBRecordReader paging query is incorrect.
> Line 92 should read:
>   query.append(" ) WHERE dbif_rno > ").append(split.getStart());
> instead of (note > instead of >=)
>   query.append(" ) WHERE dbif_rno >= ").append(split.getStart());
> Otherwise some rows will be ignored and some counted more than once.
> A map/reduce job that counts the number of rows based on a predicate will highlight the incorrect behavior.



--
This message was sent by Atlassian JIRA
(v6.2#6252)