You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@seatunnel.apache.org by GitBox <gi...@apache.org> on 2022/06/29 10:35:41 UTC
[GitHub] [incubator-seatunnel] Bingz2 opened a new issue, #2084: [Bug] [Spark Connector]Reading hive using Spark JDBC Source fails to return the correct data
Bingz2 opened a new issue, #2084:
URL: https://github.com/apache/incubator-seatunnel/issues/2084
### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/incubator-seatunnel/issues?q=is%3Aissue+label%3A%22bug%22) and found no similar issues.
### What happened
Reading a hive using Spark JDBC Source fails to return the correct data, which is actually column names rather than actual data
### SeaTunnel Version
2.1.2
### SeaTunnel Config
```conf
source {
# This is a example input plugin **only for test and demonstrate the feature input plugin**
jdbc {
driver = "org.apache.hive.jdbc.HiveDriver"
url = "jdbc:hive2://ip:10000/dws"
table = "(select mac,user_id,province_code from dws.dws_gvp_user_mac_state_stat_dd where day='2022-06-01') tmp"
result_table_name = "tmp"
user = "hive"
password = "hive"
jdbc.fetchsize = 10000
}
}
transform {
# you can also use other filter plugins, such as sql
# sql {
# sql = "select * from accesslog where request_time > 1000"
# }
}
sink {
console {
limit = 10,
serializer = "json"
}
}
```
### Running Command
```shell
./bin/start-seatunnel-spark.sh -m yarn -e client -c ./config/spark.batch.hive.jdbc.conf
```
### Error Exception
```log
22/06/28 17:39:12 INFO scheduler.DAGScheduler: Missing parents: List()
22/06/28 17:39:12 INFO scheduler.DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[6] at take at Console.scala:47), which has no missing parents
22/06/28 17:39:13 INFO memory.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 18.3 KB, free 366.3 MB)
22/06/28 17:39:13 INFO memory.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 8.4 KB, free 366.3 MB)
22/06/28 17:39:13 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on slave1.test.gitv.we:16907 (size: 8.4 KB, free: 366.3 MB)
22/06/28 17:39:13 INFO spark.SparkContext: Created broadcast 0 from broadcast at DAGScheduler.scala:1161
22/06/28 17:39:13 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (MapPartitionsRDD[6] at take at Console.scala:47) (first 15 tasks are for partitions Vector(0))
22/06/28 17:39:13 INFO cluster.YarnScheduler: Adding task set 0.0 with 1 tasks
22/06/28 17:39:13 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
22/06/28 17:39:13 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, slave4.test.gitv.we, executor 1, partition 0, PROCESS_LOCAL, 7701 bytes)
22/06/28 17:39:14 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on slave4.test.gitv.we:39237 (size: 8.4 KB, free: 912.3 MB)
22/06/28 17:39:17 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 4138 ms on slave4.test.gitv.we (executor 1) (1/1)
22/06/28 17:39:17 INFO cluster.YarnScheduler: Removed TaskSet 0.0, whose tasks have all completed, from pool
22/06/28 17:39:17 INFO scheduler.DAGScheduler: ResultStage 0 (take at Console.scala:47) finished in 4.616 s
22/06/28 17:39:17 INFO scheduler.DAGScheduler: Job 0 finished: take at Console.scala:47, took 4.674037 s
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
{"mac":"mac","user_id":"user_id","province_code":"province_code"}
22/06/28 17:39:17 INFO spark.SparkContext: Invoking stop() from shutdown hook
22/06/28 17:39:17 INFO server.AbstractConnector: Stopped Spark@4e628b52{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
22/06/28 17:39:17 INFO ui.SparkUI: Stopped Spark web UI at http://slave1.test.gitv.we:4040
22/06/28 17:39:17 INFO cluster.YarnClientSchedulerBackend: Interrupting monitor thread
22/06/28 17:39:17 INFO cluster.YarnClientSchedulerBackend: Shutting down all executors
22/06/28 17:39:17 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: Asking each executor to shut down
22/06/28 17:39:17 INFO cluster.SchedulerExtensionServices: Stopping SchedulerExtensionServices
(serviceOption=None,
services=List(),
started=false)
22/06/28 17:39:17 INFO cluster.YarnClientSchedulerBackend: Stopped
22/06/28 17:39:17 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
22/06/28 17:39:17 INFO memory.MemoryStore: MemoryStore cleared
22/06/28 17:39:17 INFO storage.BlockManager: BlockManager stopped
22/06/28 17:39:17 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
22/06/28 17:39:17 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
22/06/28 17:39:17 INFO spark.SparkContext: Successfully stopped SparkContext
22/06/28 17:39:17 INFO util.ShutdownHookManager: Shutdown hook called
```
### Flink or Spark Version
Spark version 2.4.0.cloudera2
### Java or Scala Version
Scala version 2.11.12
java version 1.8.0_112
### Screenshots
![无标题](https://user-images.githubusercontent.com/32196893/176331805-e91aed90-4fef-4044-9b34-44a008300194.png)
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@seatunnel.apache.org.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org