You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Sunita Arvind <su...@gmail.com> on 2016/06/02 17:47:26 UTC

Seeking advice on realtime querying over JDBC

Hi Experts,

We are trying to get a kafka stream ingested in Spark and expose the
registered table over JDBC for querying. Here are some questions:
1. Spark Streaming supports single context per application right? If I have
multiple customers and would like to create a kafka topic for each of them
and 1 streaming context for every topic is this doable? As per the current
spark documentation,
http://spark.apache.org/docs/latest/streaming-programming-guide.html#initializing-streamingcontext
I can have only 1 active streaming context at a time. Is there no way
around that? The use case here is, if I am looking at a 5 min window, the
window should have records for that customer only, which is possible only
by having customer specific streaming context.

2. If I am able to create multiple contexts in this fashion, can I register
them as temp tables in my application and expose them over JDBC. Going by
https://forums.databricks.com/questions/1464/how-to-configure-thrift-server-to-use-a-custom-spa.html,
looks like I can connect the thrift server to a single sparkSQL Context.
Having multiple streaming contexts means I automatically have multiple SQL
contexts?

3. Can I use SQLContext or do I need to have HiveContext in order to see
the tables registered via Spark application through the JDBC?

regards
Sunita

Re: Seeking advice on realtime querying over JDBC

Posted by Mich Talebzadeh <mi...@gmail.com>.
what is the source of your data? is that an RDMS database plus the topics
streamed via Kafka from other sources?




Dr Mich Talebzadeh



LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 2 June 2016 at 18:58, Cody Koeninger <co...@koeninger.org> wrote:

> Why are you wanting to expose spark over jdbc as opposed to just
> inserting the records from kafka into a jdbc compatible data store?
>
> On Thu, Jun 2, 2016 at 12:47 PM, Sunita Arvind <su...@gmail.com>
> wrote:
> > Hi Experts,
> >
> > We are trying to get a kafka stream ingested in Spark and expose the
> > registered table over JDBC for querying. Here are some questions:
> > 1. Spark Streaming supports single context per application right? If I
> have
> > multiple customers and would like to create a kafka topic for each of
> them
> > and 1 streaming context for every topic is this doable? As per the
> current
> > spark documentation,
> >
> http://spark.apache.org/docs/latest/streaming-programming-guide.html#initializing-streamingcontext
> > I can have only 1 active streaming context at a time. Is there no way
> around
> > that? The use case here is, if I am looking at a 5 min window, the window
> > should have records for that customer only, which is possible only by
> having
> > customer specific streaming context.
> >
> > 2. If I am able to create multiple contexts in this fashion, can I
> register
> > them as temp tables in my application and expose them over JDBC. Going by
> >
> https://forums.databricks.com/questions/1464/how-to-configure-thrift-server-to-use-a-custom-spa.html
> ,
> > looks like I can connect the thrift server to a single sparkSQL Context.
> > Having multiple streaming contexts means I automatically have multiple
> SQL
> > contexts?
> >
> > 3. Can I use SQLContext or do I need to have HiveContext in order to see
> the
> > tables registered via Spark application through the JDBC?
> >
> > regards
> > Sunita
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
> For additional commands, e-mail: user-help@spark.apache.org
>
>

Re: Seeking advice on realtime querying over JDBC

Posted by Cody Koeninger <co...@koeninger.org>.
Why are you wanting to expose spark over jdbc as opposed to just
inserting the records from kafka into a jdbc compatible data store?

On Thu, Jun 2, 2016 at 12:47 PM, Sunita Arvind <su...@gmail.com> wrote:
> Hi Experts,
>
> We are trying to get a kafka stream ingested in Spark and expose the
> registered table over JDBC for querying. Here are some questions:
> 1. Spark Streaming supports single context per application right? If I have
> multiple customers and would like to create a kafka topic for each of them
> and 1 streaming context for every topic is this doable? As per the current
> spark documentation,
> http://spark.apache.org/docs/latest/streaming-programming-guide.html#initializing-streamingcontext
> I can have only 1 active streaming context at a time. Is there no way around
> that? The use case here is, if I am looking at a 5 min window, the window
> should have records for that customer only, which is possible only by having
> customer specific streaming context.
>
> 2. If I am able to create multiple contexts in this fashion, can I register
> them as temp tables in my application and expose them over JDBC. Going by
> https://forums.databricks.com/questions/1464/how-to-configure-thrift-server-to-use-a-custom-spa.html,
> looks like I can connect the thrift server to a single sparkSQL Context.
> Having multiple streaming contexts means I automatically have multiple SQL
> contexts?
>
> 3. Can I use SQLContext or do I need to have HiveContext in order to see the
> tables registered via Spark application through the JDBC?
>
> regards
> Sunita

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org