You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "Prasanth Jayachandran (JIRA)" <ji...@apache.org> on 2016/05/03 01:10:12 UTC

[jira] [Commented] (HIVE-13546) Patch for HIVE-12893 is broken in branch-1

    [ https://issues.apache.org/jira/browse/HIVE-13546?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15267703#comment-15267703 ] 

Prasanth Jayachandran commented on HIVE-13546:
----------------------------------------------

+1. [~nemon] Thanks for looking into this!

> Patch for HIVE-12893 is broken in branch-1 
> -------------------------------------------
>
>                 Key: HIVE-13546
>                 URL: https://issues.apache.org/jira/browse/HIVE-13546
>             Project: Hive
>          Issue Type: Bug
>          Components: Logical Optimizer
>    Affects Versions: 1.3.0
>            Reporter: Nemon Lou
>         Attachments: HIVE-13546.patch
>
>
> The following sql fails:
> {noformat}
> set hive.map.aggr=true;
> set mapreduce.reduce.speculative=false;
> set hive.auto.convert.join=true;
> set hive.optimize.reducededuplication = false;
> set hive.optimize.reducededuplication.min.reducer=1;
> set hive.optimize.mapjoin.mapreduce=true;
> set hive.stats.autogather=true;
> set mapred.reduce.parallel.copies=30;
> set mapred.job.shuffle.input.buffer.percent=0.5;
> set mapred.job.reduce.input.buffer.percent=0.2;
> set mapred.map.child.java.opts=-server -Xmx2800m -Djava.net.preferIPv4Stack=true;
> set mapred.reduce.child.java.opts=-server -Xmx3800m -Djava.net.preferIPv4Stack=true;
> set mapreduce.map.memory.mb=3072;
> set mapreduce.reduce.memory.mb=4096;
> set hive.enforce.bucketing=true;
> set hive.enforce.sorting=true;
> set hive.exec.dynamic.partition.mode=nonstrict;
> set hive.exec.max.dynamic.partitions.pernode=100000;
> set hive.exec.max.dynamic.partitions=100000;
> set hive.exec.max.created.files=1000000;
> set hive.exec.parallel=true;
> set hive.exec.reducers.max=2000;
> set hive.stats.autogather=true;
> set hive.optimize.sort.dynamic.partition=true;
> set mapred.job.reduce.input.buffer.percent=0.0;
> set mapreduce.input.fileinputformat.split.minsizee=240000000;
> set mapreduce.input.fileinputformat.split.minsize.per.node=240000000;
> set mapreduce.input.fileinputformat.split.minsize.per.rack=240000000;
> set hive.optimize.sort.dynamic.partition=true;
> use tpcds_bin_partitioned_orc_4;
> insert overwrite table store_sales partition (ss_sold_date_sk)
> select
>         ss.ss_sold_time_sk,
>         ss.ss_item_sk,
>         ss.ss_customer_sk,
>         ss.ss_cdemo_sk,
>         ss.ss_hdemo_sk,
>         ss.ss_addr_sk,
>         ss.ss_store_sk,
>         ss.ss_promo_sk,
>         ss.ss_ticket_number,
>         ss.ss_quantity,
>         ss.ss_wholesale_cost,
>         ss.ss_list_price,
>         ss.ss_sales_price,
>         ss.ss_ext_discount_amt,
>         ss.ss_ext_sales_price,
>         ss.ss_ext_wholesale_cost,
>         ss.ss_ext_list_price,
>         ss.ss_ext_tax,
>         ss.ss_coupon_amt,
>         ss.ss_net_paid,
>         ss.ss_net_paid_inc_tax,
>         ss.ss_net_profit,
>         ss.ss_sold_date_sk
>       from tpcds_text_4.store_sales ss;
> {noformat}
> Error log is as follows
> {noformat}
> 2016-04-19 15:15:35,252 FATAL [main] ExecReducer: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row (tag=0) {"key":{"reducesinkkey0":null},"value":{"_col0":null,"_col1":5588,"_col2":170300,"_col3":null,"_col4":756,"_col5":91384,"_col6":16,"_col7":null,"_col8":855582,"_col9":28,"_col10":null,"_col11":48.83,"_col12":null,"_col13":0.0,"_col14":null,"_col15":899.64,"_col16":null,"_col17":6.14,"_col18":0.0,"_col19":null,"_col20":null,"_col21":null,"_col22":null}}
> 	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:244)
> 	at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
> 	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
> 	at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:180)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:422)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1732)
> 	at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:174)
> Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
> 	at java.util.ArrayList.rangeCheck(ArrayList.java:653)
> 	at java.util.ArrayList.get(ArrayList.java:429)
> 	at org.apache.hadoop.hive.common.FileUtils.makePartName(FileUtils.java:151)
> 	at org.apache.hadoop.hive.common.FileUtils.makePartName(FileUtils.java:131)
> 	at org.apache.hadoop.hive.ql.exec.FileSinkOperator.getDynPartDirectory(FileSinkOperator.java:1003)
> 	at org.apache.hadoop.hive.ql.exec.FileSinkOperator.getDynOutPaths(FileSinkOperator.java:919)
> 	at org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:713)
> 	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:235)
> 	... 7 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)