You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "lkl (Jira)" <ji...@apache.org> on 2022/02/25 12:53:00 UTC
[jira] [Updated] (HIVE-25984) when set hive.auto.convert.join=true; and set hive.exec.parallel=true; in the case cause error
[ https://issues.apache.org/jira/browse/HIVE-25984?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
lkl updated HIVE-25984:
-----------------------
Description:
{code:java}
> set hive.exec.parallel=true;
hive> set hive.exec.parallel.thread.number=16;
Query ID = hadoop_20220225202936_1afb51d0-ce67-4bc2-9794-8c82b32efe99
Total jobs = 11
Launching Job 1 out of 11
Launching Job 2 out of 11
Launching Job 3 out of 11
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
Number of reduce tasks not specified. Estimated from input data size: 1
set hive.exec.reducers.max=<number>
In order to change the average load for a reducer (in bytes):
In order to set a constant number of reducers:
set hive.exec.reducers.bytes.per.reducer=<number>
set mapreduce.job.reduces=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Launching Job 4 out of 11
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Starting Job = job_1645755235953_36462, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36462/
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36462
Starting Job = job_1645755235953_36460, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36460/
Starting Job = job_1645755235953_36463, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36463/
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36460
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36463
Starting Job = job_1645755235953_36461, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36461/
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36461
Hadoop job information for Stage-3: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:43,598 Stage-3 map = 0%, reduce = 0%
Hadoop job information for Stage-9: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:43,634 Stage-9 map = 0%, reduce = 0%
Hadoop job information for Stage-7: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:43,658 Stage-7 map = 0%, reduce = 0%
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:44,646 Stage-1 map = 0%, reduce = 0%
2022-02-25 20:29:51,767 Stage-9 map = 100%, reduce = 0%, Cumulative CPU 5.29 sec
2022-02-25 20:29:51,782 Stage-7 map = 100%, reduce = 0%, Cumulative CPU 5.45 sec
2022-02-25 20:29:52,750 Stage-3 map = 100%, reduce = 0%, Cumulative CPU 6.06 sec
2022-02-25 20:29:54,835 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 7.76 sec
2022-02-25 20:29:58,872 Stage-9 map = 100%, reduce = 100%, Cumulative CPU 7.49 sec
2022-02-25 20:29:58,883 Stage-7 map = 100%, reduce = 100%, Cumulative CPU 8.86 sec
2022-02-25 20:29:59,868 Stage-3 map = 100%, reduce = 100%, Cumulative CPU 9.96 sec
MapReduce Total cumulative CPU time: 7 seconds 490 msec
Ended Job = job_1645755235953_36463
MapReduce Total cumulative CPU time: 8 seconds 860 msec
Ended Job = job_1645755235953_36461
Stage-15 is selected by condition resolver.
Stage-8 is filtered out by condition resolver.
MapReduce Total cumulative CPU time: 9 seconds 960 msec
Ended Job = job_1645755235953_36462
Launching Job 6 out of 11
FAILED: Hive Internal Error: java.util.ConcurrentModificationException(null)
java.util.ConcurrentModificationException
at java.util.Hashtable$Enumerator.next(Hashtable.java:1387)
at org.apache.hadoop.conf.Configuration.iterator(Configuration.java:2910)
at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.initialize(ExecDriver.java:178)
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2649)
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:323)
at org.apache.hadoop.util.RunJar.main(RunJar.java:236)MapReduce Jobs Launched:
Stage-Stage-9: Map: 1 Reduce: 1 Cumulative CPU: 7.49 sec HDFS Read: 24106 HDFS Write: 1454898 SUCCESS
Stage-Stage-7: Map: 1 Reduce: 1 Cumulative CPU: 8.86 sec HDFS Read: 24150 HDFS Write: 2500859 SUCCESS
Stage-Stage-3: Map: 1 Reduce: 1 Cumulative CPU: 9.96 sec HDFS Read: 24542 HDFS Write: 10738925 SUCCESS
Total MapReduce CPU Time Spent: 26 seconds 310 msec
2022-02-25 20:30:02,955 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 7.76 sec
MapReduce Total cumulative CPU time: 7 seconds 760 msec
SLF4J: Found binding in [jar:file:/usr/local/service/hive/lib/log4j-slf4j-impl-2.17.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/local/service/hive/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/local/service/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
Execution failed with exit status: -101
Obtaining error informationTask failed!
Task ID:
Stage-15Logs:org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.lang.InterruptedException: sleep interrupted
at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:350)
at org.apache.hadoop.mapred.ClientServiceDelegate.getTaskCompletionEvents(ClientServiceDelegate.java:398)
at org.apache.hadoop.mapred.YARNRunner.getTaskCompletionEvents(YARNRunner.java:879)
at org.apache.hadoop.mapreduce.Job$6.run(Job.java:732)
at org.apache.hadoop.mapreduce.Job$6.run(Job.java:729)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1762)
at org.apache.hadoop.mapreduce.Job.getTaskCompletionEvents(Job.java:729)
at org.apache.hadoop.mapred.JobClient$NetworkedJob.getTaskCompletionEvents(JobClient.java:355)
at org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.computeReducerTimeStatsPerJob(HadoopJobExecHelper.java:612)
at org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.progress(HadoopJobExecHelper.java:570)
at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:433)
at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:149)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:76)
Caused by: java.lang.InterruptedException: sleep interrupted
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:347)
... 16 more
Ended Job = job_1645755235953_36460 with exception 'org.apache.hadoop.yarn.exceptions.YarnRuntimeException(java.lang.InterruptedException: sleep interrupted)'
hive>
> java.io.IOException: Stream closed
at java.io.BufferedInputStream.getBufIfOpen(BufferedInputStream.java:170)
at java.io.BufferedInputStream.read(BufferedInputStream.java:336)
at sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284)
at sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326)
at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178)
at java.io.InputStreamReader.read(InputStreamReader.java:184)
at java.io.BufferedReader.fill(BufferedReader.java:161)
at java.io.BufferedReader.readLine(BufferedReader.java:324)
at java.io.BufferedReader.readLine(BufferedReader.java:389)
at org.apache.hive.common.util.StreamPrinter.run(StreamPrinter.java:58)
Exception in thread "Thread-222-LocalTask-MAPREDLOCAL-stderr" java.lang.NullPointerException
at org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask.lambda$executeInChildVM$0(MapredLocalTask.java:330)
at org.apache.hadoop.hive.common.log.LogRedirector.run(LogRedirector.java:73)
at java.lang.Thread.run(Thread.java:748)
{code}
was:
{code:java}
> set hive.exec.parallel=true;
hive> set hive.exec.parallel.thread.number=16;
hive> ADD JAR ofs://f4muzj1eelr-SyDy.chdfs.ap-beijing.myqcloud.com/datam/dota-archive-ningxia/dota/emr-steps/bigdata-dw-udf-0.0.1-SNAPSHOT-jar-with-dependencies.jar;
Added [/data/emr/hive/tmp/2fbfd169-5bd0-4a63-922a-a25e88737375_resources/bigdata-dw-udf-0.0.1-SNAPSHOT-jar-with-dependencies.jar] to class path
Added resources: [ofs://f4muzj1eelr-SyDy.chdfs.ap-beijing.myqcloud.com/datam/dota-archive-ningxia/dota/emr-steps/bigdata-dw-udf-0.0.1-SNAPSHOT-jar-with-dependencies.jar]
hive>
> --INSERT OVERWRITE TABLE mgdm.dm_log_weixin_sdk_playtime_hour PARTITION(pday=20220212,phour='08',pbid='weixin')
> select
> a.ip as ip, -- ip
> a.isp_id as isp_id, -- 运营商ID
> a.isp as isp, -- 运营商名称
> a.country_id as country_id, -- 国家id
> a.country as country, -- 国家名称
> a.is_domestic as is_domestic, --
> a.province_id as province_id, -- 省份ID
> a.province as province, -- 省份名称
> a.city_id as city_id, -- 城市ID
> a.city as city, -- 城市名称
> a.did ,
> a.sessionid ,
> a.uuid ,
> a.uvip ,
> a.url ,
> a.ver ,
> a.suuid ,
> a.termid ,
> a.pix ,
> a.bid ,
> a.sdkver ,
> a.`from` ,
> a.pay ,
> a.pt ,
> a.cpt ,
> a.plid ,
> a.istry ,
> a.def ,
> a.ap ,
> a.pstatus ,
> a.cdnip ,
> a.cp ,
> a.bdid ,
> a.bsid ,
> a.cf ,
> a.cid ,
> a.idx ,
> a.vts ,
> a.td ,
> a.unionid ,
> a.src ,
> a.ct ,
> a.ht ,
> a.clip_id ,
> a.part_id ,
> a.class_id ,
> a.is_full ,
> a.duration ,
> IF(b.play_time>4000, 4000, IF(b.play_time > 0, b.play_time, 0))
> as playtime, -- 播放时长
> current_timestamp() as fetl_time -- etl时间
> from (select a.*
> from (select a.*
> from (select a.*,
> row_number() over(partition by suuid, pday, phour order by event_time desc) rn
> from mgdw.dw_log_weixin_sdk_hb_hour a
> where pday = 20220212
> and phour = '08'
> and pbid = 'weixin'
> and suuid is not null
> and logtype='hb') a
> where rn = 1) a) a
> left join (select a.pday,
> a.phour,
> a.suuid,
> ceil(a.play_hb_time - coalesce(buffer_play_time, 0)) as play_time
> from (select a.pday,
> a.phour,
> a.suuid,
> sum(play_hb_time) as play_hb_time
> from (select a.pday,
> a.phour,
> a.suuid,
> case
> when idx = min_idx then
> if(unix_timestamp(event_time) -
> unix_timestamp(min_stime) > hb_time,
> hb_time,
> unix_timestamp(event_time) -
> unix_timestamp(min_stime))
> when idx = max_idx then
> if(unix_timestamp(event_time) -
> unix_timestamp(pre_time) > hb_time,
> hb_time,
> unix_timestamp(event_time) -
> unix_timestamp(pre_time))
> else
> hb_time
> end play_hb_time
> from (select suuid, -- suuid
> idx as idx, -- 心跳序号
> event_time, -- 事件事件
> pday,
> phour,
> concat(substr(event_time, 1, 13),':00:00') as min_stime, -- 当前时段最小
> lag(event_time, 1) over(partition by suuid order by event_time) as pre_time, -- 前一个事件时间
> case when idx = 0 then 3
> when idx = 1 then 2
> when idx = 2 then 10
> when idx = 3 then 30
> when idx = 4 then 15
> when idx < 0 then 0
> when idx is null then 0
> else 120
> end as hb_time,
> min(cast(idx as int)) over(partition by suuid) as min_idx, -- 本时段最小上报
> max(cast(idx as int)) over(partition by suuid) as max_idx -- 本时段最大上报
> from mgdw.dw_log_weixin_sdk_hb_hour a
> where a.pday = 20220212
> and phour = '08'
> and pbid = 'weixin'
> and suuid is not null) a) a
> group by a.pday, a.phour, a.suuid) a -- hb心跳计算逻辑
> left join (select a.suuid,
> a.pday,
> a.phour,
> sum(buffer_play_time) / 1000 as buffer_play_time -- buffer时间
> from (select a.suuid,
> a.pday,
> a.phour,
> case
> when a.first_idx = a.idx then -- 首次buffer
> if((unix_timestamp(a.event_time) -
> unix_timestamp(a.min_stime)) * 1000 >
> a.buffer_time,
> a.buffer_time,
> (unix_timestamp(a.event_time) -
> unix_timestamp(a.min_stime)) * 1000)
> when b.suuid is not null and
> unix_timestamp(b.last_stime) -
> unix_timestamp(a.event_time) > 0 then -- 退出
> a.buffer_time
> when b.suuid is null and
> unix_timestamp(a.max_stime) -
> unix_timestamp(a.event_time) > 0 then -- 没有退出
> a.buffer_time
> else
> 0
> end as buffer_play_time
> from (select pday,
> phour,
> suuid, -- suuid
> idx, -- 心跳序号
> event_time, -- 事件事件
> concat(substr(event_time,1,13),':00:00') as min_stime, -- 当前时段最小
> concat(substr(event_time,1,13),':59:59') as max_stime, -- 当前最大时间
> 0 as buffer_time,
> min(cast(idx as int)) over(partition by suuid) as first_idx -- 本小时buffer事件的第一次上报批次号
> from mgdw.dw_log_weixin_sdk_hb_hour a
> where a.pday = 20220212
> and phour = '08'
> and pbid = 'weixin'
> and logtype='buffer') a
> left join (select a.*
> from (select pday,
> phour,
> suuid,
> event_time as last_stime,
> row_number() over(partition by suuid order by event_time desc) rn
> from mgdw.dw_log_weixin_sdk_hb_hour
> where pday = 20220212
> and phour = '08'
> and pbid = 'weixin'
> and suuid is not null
> and ht = 2
> and logtype='hb') a
> where rn = 1) b
> on a.pday = b.pday
> and a.phour = b.phour
> and a.suuid = b.suuid) a
> group by a.suuid, a.pday, a.phour) b -- buffer计算逻辑
> on a.suuid = b.suuid
> and a.pday = b.pday
> and a.phour = b.phour) b
> on a.pday = b.pday
> and a.phour = a.phour
> and a.suuid = b.suuid;
Query ID = hadoop_20220225202936_1afb51d0-ce67-4bc2-9794-8c82b32efe99
Total jobs = 11
Launching Job 1 out of 11
Launching Job 2 out of 11
Launching Job 3 out of 11
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
Number of reduce tasks not specified. Estimated from input data size: 1
set hive.exec.reducers.max=<number>
In order to change the average load for a reducer (in bytes):
In order to set a constant number of reducers:
set hive.exec.reducers.bytes.per.reducer=<number>
set mapreduce.job.reduces=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Launching Job 4 out of 11
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Starting Job = job_1645755235953_36462, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36462/
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36462
Starting Job = job_1645755235953_36460, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36460/
Starting Job = job_1645755235953_36463, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36463/
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36460
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36463
Starting Job = job_1645755235953_36461, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36461/
Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36461
Hadoop job information for Stage-3: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:43,598 Stage-3 map = 0%, reduce = 0%
Hadoop job information for Stage-9: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:43,634 Stage-9 map = 0%, reduce = 0%
Hadoop job information for Stage-7: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:43,658 Stage-7 map = 0%, reduce = 0%
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1
2022-02-25 20:29:44,646 Stage-1 map = 0%, reduce = 0%
2022-02-25 20:29:51,767 Stage-9 map = 100%, reduce = 0%, Cumulative CPU 5.29 sec
2022-02-25 20:29:51,782 Stage-7 map = 100%, reduce = 0%, Cumulative CPU 5.45 sec
2022-02-25 20:29:52,750 Stage-3 map = 100%, reduce = 0%, Cumulative CPU 6.06 sec
2022-02-25 20:29:54,835 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 7.76 sec
2022-02-25 20:29:58,872 Stage-9 map = 100%, reduce = 100%, Cumulative CPU 7.49 sec
2022-02-25 20:29:58,883 Stage-7 map = 100%, reduce = 100%, Cumulative CPU 8.86 sec
2022-02-25 20:29:59,868 Stage-3 map = 100%, reduce = 100%, Cumulative CPU 9.96 sec
MapReduce Total cumulative CPU time: 7 seconds 490 msec
Ended Job = job_1645755235953_36463
MapReduce Total cumulative CPU time: 8 seconds 860 msec
Ended Job = job_1645755235953_36461
Stage-15 is selected by condition resolver.
Stage-8 is filtered out by condition resolver.
MapReduce Total cumulative CPU time: 9 seconds 960 msec
Ended Job = job_1645755235953_36462
Launching Job 6 out of 11
FAILED: Hive Internal Error: java.util.ConcurrentModificationException(null)
java.util.ConcurrentModificationException
at java.util.Hashtable$Enumerator.next(Hashtable.java:1387)
at org.apache.hadoop.conf.Configuration.iterator(Configuration.java:2910)
at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.initialize(ExecDriver.java:178)
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2649)
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:323)
at org.apache.hadoop.util.RunJar.main(RunJar.java:236)MapReduce Jobs Launched:
Stage-Stage-9: Map: 1 Reduce: 1 Cumulative CPU: 7.49 sec HDFS Read: 24106 HDFS Write: 1454898 SUCCESS
Stage-Stage-7: Map: 1 Reduce: 1 Cumulative CPU: 8.86 sec HDFS Read: 24150 HDFS Write: 2500859 SUCCESS
Stage-Stage-3: Map: 1 Reduce: 1 Cumulative CPU: 9.96 sec HDFS Read: 24542 HDFS Write: 10738925 SUCCESS
Total MapReduce CPU Time Spent: 26 seconds 310 msec
2022-02-25 20:30:02,955 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 7.76 sec
MapReduce Total cumulative CPU time: 7 seconds 760 msec
SLF4J: Found binding in [jar:file:/usr/local/service/hive/lib/log4j-slf4j-impl-2.17.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/local/service/hive/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/local/service/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
Execution failed with exit status: -101
Obtaining error informationTask failed!
Task ID:
Stage-15Logs:org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.lang.InterruptedException: sleep interrupted
at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:350)
at org.apache.hadoop.mapred.ClientServiceDelegate.getTaskCompletionEvents(ClientServiceDelegate.java:398)
at org.apache.hadoop.mapred.YARNRunner.getTaskCompletionEvents(YARNRunner.java:879)
at org.apache.hadoop.mapreduce.Job$6.run(Job.java:732)
at org.apache.hadoop.mapreduce.Job$6.run(Job.java:729)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1762)
at org.apache.hadoop.mapreduce.Job.getTaskCompletionEvents(Job.java:729)
at org.apache.hadoop.mapred.JobClient$NetworkedJob.getTaskCompletionEvents(JobClient.java:355)
at org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.computeReducerTimeStatsPerJob(HadoopJobExecHelper.java:612)
at org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.progress(HadoopJobExecHelper.java:570)
at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:433)
at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:149)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:76)
Caused by: java.lang.InterruptedException: sleep interrupted
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:347)
... 16 more
Ended Job = job_1645755235953_36460 with exception 'org.apache.hadoop.yarn.exceptions.YarnRuntimeException(java.lang.InterruptedException: sleep interrupted)'
hive>
> java.io.IOException: Stream closed
at java.io.BufferedInputStream.getBufIfOpen(BufferedInputStream.java:170)
at java.io.BufferedInputStream.read(BufferedInputStream.java:336)
at sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284)
at sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326)
at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178)
at java.io.InputStreamReader.read(InputStreamReader.java:184)
at java.io.BufferedReader.fill(BufferedReader.java:161)
at java.io.BufferedReader.readLine(BufferedReader.java:324)
at java.io.BufferedReader.readLine(BufferedReader.java:389)
at org.apache.hive.common.util.StreamPrinter.run(StreamPrinter.java:58)
Exception in thread "Thread-222-LocalTask-MAPREDLOCAL-stderr" java.lang.NullPointerException
at org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask.lambda$executeInChildVM$0(MapredLocalTask.java:330)
at org.apache.hadoop.hive.common.log.LogRedirector.run(LogRedirector.java:73)
at java.lang.Thread.run(Thread.java:748)
{code}
> when set hive.auto.convert.join=true; and set hive.exec.parallel=true; in the case cause error
> ----------------------------------------------------------------------------------------------
>
> Key: HIVE-25984
> URL: https://issues.apache.org/jira/browse/HIVE-25984
> Project: Hive
> Issue Type: Improvement
> Components: Hive
> Affects Versions: 3.0.0, 3.1.1, 3.1.2
> Reporter: lkl
> Priority: Major
>
> {code:java}
> > set hive.exec.parallel=true;
> hive> set hive.exec.parallel.thread.number=16;
> Query ID = hadoop_20220225202936_1afb51d0-ce67-4bc2-9794-8c82b32efe99
> Total jobs = 11
> Launching Job 1 out of 11
> Launching Job 2 out of 11
> Launching Job 3 out of 11
> Number of reduce tasks not specified. Estimated from input data size: 1
> In order to change the average load for a reducer (in bytes):
> set hive.exec.reducers.bytes.per.reducer=<number>
> In order to limit the maximum number of reducers:
> Number of reduce tasks not specified. Estimated from input data size: 1
> set hive.exec.reducers.max=<number>
> In order to change the average load for a reducer (in bytes):
> In order to set a constant number of reducers:
> set hive.exec.reducers.bytes.per.reducer=<number>
> set mapreduce.job.reduces=<number>
> In order to limit the maximum number of reducers:
> set hive.exec.reducers.max=<number>
> In order to set a constant number of reducers:
> set mapreduce.job.reduces=<number>
> Launching Job 4 out of 11
> Number of reduce tasks not specified. Estimated from input data size: 1
> In order to change the average load for a reducer (in bytes):
> set hive.exec.reducers.bytes.per.reducer=<number>
> In order to limit the maximum number of reducers:
> set hive.exec.reducers.max=<number>
> In order to set a constant number of reducers:
> set mapreduce.job.reduces=<number>
> Number of reduce tasks not specified. Estimated from input data size: 1
> In order to change the average load for a reducer (in bytes):
> set hive.exec.reducers.bytes.per.reducer=<number>
> In order to limit the maximum number of reducers:
> set hive.exec.reducers.max=<number>
> In order to set a constant number of reducers:
> set mapreduce.job.reduces=<number>
> Starting Job = job_1645755235953_36462, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36462/
> Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36462
> Starting Job = job_1645755235953_36460, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36460/
> Starting Job = job_1645755235953_36463, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36463/
> Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36460
> Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36463
> Starting Job = job_1645755235953_36461, Tracking URL = http://172.21.126.228:5004/proxy/application_1645755235953_36461/
> Kill Command = /usr/local/service/hadoop/bin/mapred job -kill job_1645755235953_36461
> Hadoop job information for Stage-3: number of mappers: 1; number of reducers: 1
> 2022-02-25 20:29:43,598 Stage-3 map = 0%, reduce = 0%
> Hadoop job information for Stage-9: number of mappers: 1; number of reducers: 1
> 2022-02-25 20:29:43,634 Stage-9 map = 0%, reduce = 0%
> Hadoop job information for Stage-7: number of mappers: 1; number of reducers: 1
> 2022-02-25 20:29:43,658 Stage-7 map = 0%, reduce = 0%
> Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1
> 2022-02-25 20:29:44,646 Stage-1 map = 0%, reduce = 0%
> 2022-02-25 20:29:51,767 Stage-9 map = 100%, reduce = 0%, Cumulative CPU 5.29 sec
> 2022-02-25 20:29:51,782 Stage-7 map = 100%, reduce = 0%, Cumulative CPU 5.45 sec
> 2022-02-25 20:29:52,750 Stage-3 map = 100%, reduce = 0%, Cumulative CPU 6.06 sec
> 2022-02-25 20:29:54,835 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 7.76 sec
> 2022-02-25 20:29:58,872 Stage-9 map = 100%, reduce = 100%, Cumulative CPU 7.49 sec
> 2022-02-25 20:29:58,883 Stage-7 map = 100%, reduce = 100%, Cumulative CPU 8.86 sec
> 2022-02-25 20:29:59,868 Stage-3 map = 100%, reduce = 100%, Cumulative CPU 9.96 sec
> MapReduce Total cumulative CPU time: 7 seconds 490 msec
> Ended Job = job_1645755235953_36463
> MapReduce Total cumulative CPU time: 8 seconds 860 msec
> Ended Job = job_1645755235953_36461
> Stage-15 is selected by condition resolver.
> Stage-8 is filtered out by condition resolver.
> MapReduce Total cumulative CPU time: 9 seconds 960 msec
> Ended Job = job_1645755235953_36462
> Launching Job 6 out of 11
> FAILED: Hive Internal Error: java.util.ConcurrentModificationException(null)
> java.util.ConcurrentModificationException
> at java.util.Hashtable$Enumerator.next(Hashtable.java:1387)
> at org.apache.hadoop.conf.Configuration.iterator(Configuration.java:2910)
> at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.initialize(ExecDriver.java:178)
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2649)
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
> at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
> at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
> at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at org.apache.hadoop.util.RunJar.run(RunJar.java:323)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:236)MapReduce Jobs Launched:
> Stage-Stage-9: Map: 1 Reduce: 1 Cumulative CPU: 7.49 sec HDFS Read: 24106 HDFS Write: 1454898 SUCCESS
> Stage-Stage-7: Map: 1 Reduce: 1 Cumulative CPU: 8.86 sec HDFS Read: 24150 HDFS Write: 2500859 SUCCESS
> Stage-Stage-3: Map: 1 Reduce: 1 Cumulative CPU: 9.96 sec HDFS Read: 24542 HDFS Write: 10738925 SUCCESS
> Total MapReduce CPU Time Spent: 26 seconds 310 msec
> 2022-02-25 20:30:02,955 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 7.76 sec
> MapReduce Total cumulative CPU time: 7 seconds 760 msec
> SLF4J: Found binding in [jar:file:/usr/local/service/hive/lib/log4j-slf4j-impl-2.17.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/usr/local/service/hive/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/usr/local/service/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
> Execution failed with exit status: -101
> Obtaining error informationTask failed!
> Task ID:
> Stage-15Logs:org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.lang.InterruptedException: sleep interrupted
> at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:350)
> at org.apache.hadoop.mapred.ClientServiceDelegate.getTaskCompletionEvents(ClientServiceDelegate.java:398)
> at org.apache.hadoop.mapred.YARNRunner.getTaskCompletionEvents(YARNRunner.java:879)
> at org.apache.hadoop.mapreduce.Job$6.run(Job.java:732)
> at org.apache.hadoop.mapreduce.Job$6.run(Job.java:729)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1762)
> at org.apache.hadoop.mapreduce.Job.getTaskCompletionEvents(Job.java:729)
> at org.apache.hadoop.mapred.JobClient$NetworkedJob.getTaskCompletionEvents(JobClient.java:355)
> at org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.computeReducerTimeStatsPerJob(HadoopJobExecHelper.java:612)
> at org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.progress(HadoopJobExecHelper.java:570)
> at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:433)
> at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:149)
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
> at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
> at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:76)
> Caused by: java.lang.InterruptedException: sleep interrupted
> at java.lang.Thread.sleep(Native Method)
> at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:347)
> ... 16 more
> Ended Job = job_1645755235953_36460 with exception 'org.apache.hadoop.yarn.exceptions.YarnRuntimeException(java.lang.InterruptedException: sleep interrupted)'
> hive>
> > java.io.IOException: Stream closed
> at java.io.BufferedInputStream.getBufIfOpen(BufferedInputStream.java:170)
> at java.io.BufferedInputStream.read(BufferedInputStream.java:336)
> at sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284)
> at sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326)
> at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178)
> at java.io.InputStreamReader.read(InputStreamReader.java:184)
> at java.io.BufferedReader.fill(BufferedReader.java:161)
> at java.io.BufferedReader.readLine(BufferedReader.java:324)
> at java.io.BufferedReader.readLine(BufferedReader.java:389)
> at org.apache.hive.common.util.StreamPrinter.run(StreamPrinter.java:58)
> Exception in thread "Thread-222-LocalTask-MAPREDLOCAL-stderr" java.lang.NullPointerException
> at org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask.lambda$executeInChildVM$0(MapredLocalTask.java:330)
> at org.apache.hadoop.hive.common.log.LogRedirector.run(LogRedirector.java:73)
> at java.lang.Thread.run(Thread.java:748)
> {code}
--
This message was sent by Atlassian Jira
(v8.20.1#820001)