You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Yhzhtk (JIRA)" <ji...@apache.org> on 2014/08/26 06:30:57 UTC

[jira] [Created] (HIVE-7879) Execute the same sql first sucessful, but the second error "/ by zero"

Yhzhtk created HIVE-7879:
----------------------------

             Summary: Execute the same sql first sucessful, but the second error "/ by zero"
                 Key: HIVE-7879
                 URL: https://issues.apache.org/jira/browse/HIVE-7879
             Project: Hive
          Issue Type: Bug
         Environment: CentOS release 6.2 (Final)


Hadoop 2.2.0
Subversion https://svn.apache.org/repos/asf/hadoop/common -r 1529768
Compiled by hortonmu on 2013-10-07T06:28Z
Compiled with protoc 2.5.0
>From source with checksum 79e53ce7994d1628b240f09af91e1af4


hive-0.12.0

            Reporter: Yhzhtk


The hive sql is:
{quote}
        select working_experience,count(*) 
        from search 
        where 
            log_date = 20140825 and 
            ua not rlike '^.*(HttpClient|Jakarta|scrapy|bot|spider|wget).*$' and
             working_experience in ('不限','应届毕业生','1年以下','1-3年','3-5年','5-10年','10年以上','') 
        group by working_experience
{quote}

*Execute the sql a short time on the same environment(data not change) twice, the first is run sucessful and print the right result , but the second is error:*
    
The error is:
{quote}
    Diagnostic Messages for this Task: 
    / by zero
{quote}

*The first (successful), The whole info is:*
{quote}
> select working_experience,count(*) from search where log_date = 20140825 and ua not rlike '^.*(HttpClient|Jakarta|scrapy|bot|spider|wget).*$' and working_experience in ('不限','应届毕业生','1年以下','1-3年','3-5年','5-10年','10年以上','') group by working_experience; 
Total MapReduce jobs = 1 
Launching Job 1 out of 1 
Number of reduce tasks not specified. Estimated from input data size: 1 
In order to change the average load for a reducer (in bytes): 
set hive.exec.reducers.bytes.per.reducer=<number> 
In order to limit the maximum number of reducers: 
set hive.exec.reducers.max=<number> 
In order to set a constant number of reducers: 
set mapred.reduce.tasks=<number> 
Starting Job = job_1404899662896_1182, Tracking URL = http://rsyslog-16-3:8088/proxy/application_1404899662896_1182/ 
Kill Command = /apps/hadoop_mapreduce/hadoop-2.2.0/bin/hadoop job -kill job_1404899662896_1182 
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1 
2014-08-26 11:41:45,199 Stage-1 map = 0%, reduce = 0% 
2014-08-26 11:41:54,522 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:41:55,557 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:41:56,600 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:41:57,639 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:41:58,677 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:41:59,711 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:42:00,751 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec 
2014-08-26 11:42:02,018 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 5.91 sec 
2014-08-26 11:42:03,055 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 5.91 sec 
2014-08-26 11:42:04,099 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 5.91 sec 
MapReduce Total cumulative CPU time: 5 seconds 910 msec 
Ended Job = job_1404899662896_1182 
MapReduce Jobs Launched: 
Job 0: Map: 1 Reduce: 1 Cumulative CPU: 5.91 sec HDFS Read: 22122871 HDFS Write: 104 SUCCESS 
Total MapReduce CPU Time Spent: 5 seconds 910 msec 
OK 
50339 
1-3年 1949 
10年以上 60 
1年以下 360 
3-5年 689 
5-10年 328 
不限 1196 
应届毕业生 961 
Time taken: 26.135 seconds, Fetched: 8 row(s)
{quote}

*The second (error), The whole info is:*
{quote}
> select working_experience,count(*) from search where log_date = 20140825 and ua not rlike '^.*(HttpClient|Jakarta|scrapy|bot|spider|wget).*$' and working_experience in ('不限','应届毕业生','1年以下','1-3年','3-5年','5-10年','10年以上','') group by working_experience 
> ; 
Total MapReduce jobs = 1 
Launching Job 1 out of 1 
Number of reduce tasks not specified. Estimated from input data size: 1 
In order to change the average load for a reducer (in bytes): 
set hive.exec.reducers.bytes.per.reducer=<number> 
In order to limit the maximum number of reducers: 
set hive.exec.reducers.max=<number> 
In order to set a constant number of reducers: 
set mapred.reduce.tasks=<number> 
Starting Job = job_1404899662896_1183, Tracking URL = http://rsyslog-16-3:8088/proxy/application_1404899662896_1183/ 
Kill Command = /apps/hadoop_mapreduce/hadoop-2.2.0/bin/hadoop job -kill job_1404899662896_1183 
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1 
2014-08-26 11:42:20,923 Stage-1 map = 0%, reduce = 0% 
2014-08-26 11:42:38,491 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:39,525 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:40,563 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:41,596 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:42,644 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:43,677 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:44,712 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:45,753 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:46,786 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:47,817 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:48,859 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:49,896 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:50,929 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:51,962 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:53,000 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:54,037 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 4.26 sec 
2014-08-26 11:42:55,073 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 4.26 sec 
MapReduce Total cumulative CPU time: 4 seconds 260 msec 
Ended Job = job_1404899662896_1183 with errors 
Error during job, obtaining debugging information... 
Examining task ID: task_1404899662896_1183_m_000000 (and more) from job job_1404899662896_1183 

Task with the most failures(4): 
----- 
Task ID: 
task_1404899662896_1183_r_000000 

URL: 
http://rsyslog-16-3:8088/taskdetails.jsp?jobid=job_1404899662896_1183&tipid=task_1404899662896_1183_r_000000 
----- 
Diagnostic Messages for this Task: 
/ by zero 


FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask 
MapReduce Jobs Launched: 
Job 0: Map: 1 Reduce: 1 Cumulative CPU: 4.26 sec HDFS Read: 22122871 HDFS Write: 0 FAIL 
Total MapReduce CPU Time Spent: 4 seconds 260 msec
{quote}


Is this is a bug? How can I resolved it.

Thanks very much!



--
This message was sent by Atlassian JIRA
(v6.2#6252)