You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hive.apache.org by "Bogala, Chandra Reddy" <Ch...@gs.com> on 2014/07/11 08:55:36 UTC

beeline client

Hi,
   Currently I am submitting multiple hive jobs using hive cli with "hive -f" from different scripts. All these jobs I could see in application tracker and these get processed in parallel.
Now I planned to switch to HiveServer2 and submitting jobs using beeline client from multiple scripts  example : "nohup beeline -u jdbc:hive2://<host>:<port>#<variable>=$epoch -n <user> -p <pass> -d org.apache.hive.jdbc.HiveDriver -f <hql script>.hql &".
But I could see jobs get submitted in serial. Only one job appears in application tracker at a time even though cluster resources are available. Once that job finishes then only other job get submitted. Why is that? Is there any setting needs to be set to submit jobs in parallel?

Thanks,
Chandra


Re: beeline client

Posted by Xuefu Zhang <xz...@cloudera.com>.
Yes, it's known issue (HIVE-6758) that beeline -f  works differently when
running  in background. A workaround is noted in the JIRA.

--Xuefu


On Sun, Jul 13, 2014 at 9:43 PM, Bogala, Chandra Reddy <
Chandra.Bogala@gs.com> wrote:

> I submit 5 queries from 5 different scripts.  When I use hive cli, I could
> see all these 5 queries related jobs in application tracker in different
> states ( running/accepted etc). When I submit through beeline only one job
> I see in application tracker. Other job I can see only after first job
> finishes.
>
>
>
> Ex: hive –f script1.hql &
>
>       Hive –f script2.hql &
>
>       Hive –f script3.hql &
>
>       Hive –f script4.hql &
>
>       Hive –f script5.hql &
>
>
>
>
>
> *From:* Xuefu Zhang [mailto:xzhang@cloudera.com]
> *Sent:* Friday, July 11, 2014 7:29 PM
> *To:* user@hive.apache.org
> *Subject:* Re: beeline client
>
>
>
> Chaudra,
>
> The difference you saw between Hive CLI and Beeline might indicate a bug.
> However, before making such a conclusion, could you give an example of your
> queries? Are the jobs you expect to run parallel for a single query? Please
> note that your script file is executed line by line in either case.
>
> --Xuefu
>
>
>
> On Thu, Jul 10, 2014 at 11:55 PM, Bogala, Chandra Reddy <
> Chandra.Bogala@gs.com> wrote:
>
> Hi,
>
>    Currently I am submitting multiple hive jobs using hive cli with “hive
> –f” from different scripts. All these jobs I could see in application
> tracker and these get processed in parallel.
>
> Now I planned to switch to HiveServer2 and submitting jobs using beeline
> client from multiple scripts  example : “nohup beeline -u
> jdbc:hive2://<host>:<port>#<variable>=$epoch -n <user> -p <pass> -d
> org.apache.hive.jdbc.HiveDriver -f <hql script>.hql &”.
>
> But I could see jobs get submitted in serial. Only one job appears in
> application tracker at a time even though cluster resources are available.
> Once that job finishes then only other job get submitted. Why is that? Is
> there any setting needs to be set to submit jobs in parallel?
>
>
>
> Thanks,
>
> Chandra
>
>
>
>
>

RE: beeline client

Posted by "Bogala, Chandra Reddy" <Ch...@gs.com>.
I submit 5 queries from 5 different scripts.  When I use hive cli, I could see all these 5 queries related jobs in application tracker in different states ( running/accepted etc). When I submit through beeline only one job I see in application tracker. Other job I can see only after first job finishes.

Ex: hive –f script1.hql &
      Hive –f script2.hql &
      Hive –f script3.hql &
      Hive –f script4.hql &
      Hive –f script5.hql &


From: Xuefu Zhang [mailto:xzhang@cloudera.com]
Sent: Friday, July 11, 2014 7:29 PM
To: user@hive.apache.org
Subject: Re: beeline client

Chaudra,
The difference you saw between Hive CLI and Beeline might indicate a bug. However, before making such a conclusion, could you give an example of your queries? Are the jobs you expect to run parallel for a single query? Please note that your script file is executed line by line in either case.
--Xuefu

On Thu, Jul 10, 2014 at 11:55 PM, Bogala, Chandra Reddy <Ch...@gs.com>> wrote:
Hi,
   Currently I am submitting multiple hive jobs using hive cli with “hive –f” from different scripts. All these jobs I could see in application tracker and these get processed in parallel.
Now I planned to switch to HiveServer2 and submitting jobs using beeline client from multiple scripts  example : “nohup beeline -u jdbc:hive2://<host>:<port>#<variable>=$epoch -n <user> -p <pass> -d org.apache.hive.jdbc.HiveDriver -f <hql script>.hql &”.
But I could see jobs get submitted in serial. Only one job appears in application tracker at a time even though cluster resources are available. Once that job finishes then only other job get submitted. Why is that? Is there any setting needs to be set to submit jobs in parallel?

Thanks,
Chandra



Re: beeline client

Posted by Xuefu Zhang <xz...@cloudera.com>.
Chaudra,

The difference you saw between Hive CLI and Beeline might indicate a bug.
However, before making such a conclusion, could you give an example of your
queries? Are the jobs you expect to run parallel for a single query? Please
note that your script file is executed line by line in either case.

--Xuefu


On Thu, Jul 10, 2014 at 11:55 PM, Bogala, Chandra Reddy <
Chandra.Bogala@gs.com> wrote:

> Hi,
>
>    Currently I am submitting multiple hive jobs using hive cli with “hive
> –f” from different scripts. All these jobs I could see in application
> tracker and these get processed in parallel.
>
> Now I planned to switch to HiveServer2 and submitting jobs using beeline
> client from multiple scripts  example : “nohup beeline -u
> jdbc:hive2://<host>:<port>#<variable>=$epoch -n <user> -p <pass> -d
> org.apache.hive.jdbc.HiveDriver -f <hql script>.hql &”.
>
> But I could see jobs get submitted in serial. Only one job appears in
> application tracker at a time even though cluster resources are available.
> Once that job finishes then only other job get submitted. Why is that? Is
> there any setting needs to be set to submit jobs in parallel?
>
>
>
> Thanks,
>
> Chandra
>
>
>