You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Pravin Agrawal (JIRA)" <ji...@apache.org> on 2018/01/15 11:40:00 UTC

[jira] [Created] (SPARK-23077) Apache Structured Streaming: How to write streaming dataset into Hive?

Pravin Agrawal created SPARK-23077:
--------------------------------------

             Summary: Apache Structured Streaming: How to write streaming dataset into Hive?
                 Key: SPARK-23077
                 URL: https://issues.apache.org/jira/browse/SPARK-23077
             Project: Spark
          Issue Type: Question
          Components: Structured Streaming
    Affects Versions: 2.2.0
            Reporter: Pravin Agrawal


Using Apache Spark 2.2: Structured Streaming, I am creating a program which reads data from Kafka and write it to Hive.
I am looking for writing bulk data incoming in Kafka topic @ 100 records/sec.

**Hive Table Created:**

CREATE TABLE demo_user( timeaa BIGINT, numberbb INT, decimalcc DOUBLE, stringdd STRING, booleanee BOOLEAN ) STORED AS ORC ;


**Insert via Manual Hive Query:**

INSERT INTO TABLE demo_user (1514133139123, 14, 26.4, 'pravin', true);


**Insert via spark structured streaming code:**

SparkConf conf = new SparkConf();
 conf.setAppName("testing");
 conf.setMaster("local[2]");
 conf.set("hive.metastore.uris", "thrift://localhost:9083");
 SparkSession session = 
 SparkSession.builder().config(conf).enableHiveSupport().getOrCreate();

// workaround START: code to insert static data into hive
 String insertQuery = "INSERT INTO TABLE demo_user (1514133139123, 14, 26.4, 'pravin', true)";
 session.sql(insertQuery);
 // workaround END:

// Solution START
 Dataset<Row> dataset = readFromKafka(sparkSession); // private method reading data from Kafka's 'xyz' topic

// **My question here:**
 // some code which writes dataset into hive table demo_user
 // Solution END



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org