You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "tooptoop4 (JIRA)" <ji...@apache.org> on 2017/11/04 11:44:00 UTC
[jira] [Updated] (DRILL-5932) HiveServer2 queries throw error with
jdbc connection
[ https://issues.apache.org/jira/browse/DRILL-5932?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
tooptoop4 updated DRILL-5932:
-----------------------------
Description:
Basic hive queries all throw error!
{code:sql}
copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar
added this storage plugin:
{
"type": "jdbc",
"driver": "org.apache.hive.jdbc.HiveDriver",
"url": "jdbc:hive2://host:10000/default",
"username": "hive",
"password": "hive1234",
"enabled": true
}
[ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
[ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
apache drill 1.11.0
"this isn't your grandfather's sql"
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0
[Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)
0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
+-------+-------------------------------+
| ok | summary |
+-------+-------------------------------+
| true | exec.errors.verbose updated. |
+-------+-------------------------------+
1 row selected (0.351 seconds)
0: jdbc:drill:zk=local> select * from hive2.contact;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0
[Error Id: fe36b026-e8ff-4354-af6c-6073130680c9 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]
(org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
org.apache.hive.jdbc.Utils.verifySuccess():267
org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
org.apache.hive.jdbc.HiveStatement.execute():253
org.apache.hive.jdbc.HiveStatement.executeQuery():476
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
org.apache.drill.common.SelfCleaningRunnable.run():38
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
org.apache.hive.service.cli.operation.Operation.toSQLException():380
org.apache.hive.service.cli.operation.SQLOperation.prepare():206
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
org.apache.hadoop.hive.ql.Driver.compile():468
org.apache.hadoop.hive.ql.Driver.compileInternal():1316
org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
org.apache.hive.service.cli.operation.SQLOperation.prepare():204
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local> select * from hive2.`default`.contact;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0
[Error Id: 7b7aae92-fabd-49dd-ab84-15ca00189882 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]
(org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
org.apache.hive.jdbc.Utils.verifySuccess():267
org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
org.apache.hive.jdbc.HiveStatement.execute():253
org.apache.hive.jdbc.HiveStatement.executeQuery():476
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
org.apache.drill.common.SelfCleaningRunnable.run():38
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
org.apache.hive.service.cli.operation.Operation.toSQLException():380
org.apache.hive.service.cli.operation.SQLOperation.prepare():206
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
org.apache.hadoop.hive.ql.Driver.compile():468
org.apache.hadoop.hive.ql.Driver.compileInternal():1316
org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
org.apache.hive.service.cli.operation.SQLOperation.prepare():204
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local>
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0
[Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]
(org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
org.apache.hive.jdbc.Utils.verifySuccess():267
org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
org.apache.hive.jdbc.HiveStatement.execute():253
org.apache.hive.jdbc.HiveStatement.executeQuery():476
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
org.apache.drill.common.SelfCleaningRunnable.run():38
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
org.apache.hive.service.cli.operation.Operation.toSQLException():380
org.apache.hive.service.cli.operation.SQLOperation.prepare():206
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
org.apache.hadoop.hive.ql.Driver.compile():468
org.apache.hadoop.hive.ql.Driver.compileInternal():1316
org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
org.apache.hive.service.cli.operation.SQLOperation.prepare():204
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local>
{
"id": {
"part1": 2738948259170077000,
"part2": -61360705522985576
},
"type": 1,
"start": 1509772561834,
"end": 1509772562065,
"query": "SELECT count(*) FROM hive2.`contact`",
"plan": "00-00 Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01 Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02 Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
"foreman": {
"address": "ip-myip.mydomain.orghere.com",
"userPort": 31010,
"controlPort": 31011,
"dataPort": 31012,
"version": "1.11.0"
},
"state": 4,
"totalFragments": 1,
"finishedFragments": 0,
"fragmentProfile": [
{
"majorFragmentId": 0,
"minorFragmentProfile": [
{
"state": 0,
"minorFragmentId": 0,
"endpoint": {
"address": "ip-myip.mydomain.orghere.com",
"userPort": 31010,
"controlPort": 31011,
"dataPort": 31012,
"version": "1.11.0"
},
"lastUpdate": 1509772562054,
"lastProgress": 1509772562054
}
]
}
],
"user": "anonymous",
"error": "DATA_READ ERROR: Drill Remote Exception\n\n",
"verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
"errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
"errorNode": "ip-myip.mydomain.orghere.com:31010",
"optionsJson": "[ {\n \"kind\" : \"BOOLEAN\",\n \"type\" : \"SESSION\",\n \"name\" : \"exec.errors.verbose\",\n \"bool_val\" : true\n} ]",
"planEnd": 1509772562051,
"queueWaitEnd": 0
}
0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0
To prove that it can connect, here is the query plan, showing all columns from the table:
00-00 Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
00-01 Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
00-02 SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
00-03 Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
00-04 Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
00-05 Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
00-06 Jdbc(sql=[SELECT *
FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282
Below connection with beeline proves there is data in the table:
Beeline version 2.3.0 by Apache Hive
0: jdbc:hive2://localhost:10000/default> show create table contact;
+----------------------------------------------------+
| createtab_stmt |
+----------------------------------------------------+
| CREATE TABLE `contact`( |
| `contact_id` varchar(50), |
| `first_name` varchar(50), |
| `last_name` varchar(50), |
| `address_line_1` varchar(100), |
| `address_line_2` varchar(100), |
| `city` varchar(50), |
| `postcode` varchar(10), |
| `email` varchar(100), |
| `dob` date, |
| `gender` varchar(1), |
| `marital_status` varchar(1), |
| `tfn` varchar(20)) |
| ROW FORMAT SERDE |
| 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' |
| STORED AS INPUTFORMAT |
| 'org.apache.hadoop.mapred.TextInputFormat' |
| OUTPUTFORMAT |
| 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
| LOCATION |
| 'file:/home/ec2-user/warehouse/contact' |
| TBLPROPERTIES ( |
| 'transient_lastDdlTime'='1509487563') |
+----------------------------------------------------+
23 rows selected (0.361 seconds)
0: jdbc:hive2://localhost:10000/default> select * from contact;
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| contact.contact_id | contact.first_name | contact.last_name | contact.address_line_1 | contact.address_line_2 | contact.city | contact.postcode | contact.email | contact.dob | contact.gender | contact.marital_status | contact.tfn |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| C1 | Bob | B | St | | Ta | 2100 | bob@b.com | 2020-03-01 | M | M | 12 |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
1 rows selected (0.156 seconds)
0: jdbc:hive2://localhost:10000/default>
{code}
was:
Basic hive queries all throw error!
{code:sql}
copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar
added this storage plugin:
{
"type": "jdbc",
"driver": "org.apache.hive.jdbc.HiveDriver",
"url": "jdbc:hive2://host:10000/default",
"username": "hive",
"password": "hive1234",
"enabled": true
}
[ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
[ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
apache drill 1.11.0
"this isn't your grandfather's sql"
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0
[Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)
0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
+-------+-------------------------------+
| ok | summary |
+-------+-------------------------------+
| true | exec.errors.verbose updated. |
+-------+-------------------------------+
1 row selected (0.351 seconds)
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0
[Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]
(org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
org.apache.hive.jdbc.Utils.verifySuccess():267
org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
org.apache.hive.jdbc.HiveStatement.execute():253
org.apache.hive.jdbc.HiveStatement.executeQuery():476
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
org.apache.drill.common.SelfCleaningRunnable.run():38
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
org.apache.hive.service.cli.operation.Operation.toSQLException():380
org.apache.hive.service.cli.operation.SQLOperation.prepare():206
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
org.apache.hadoop.hive.ql.Driver.compile():468
org.apache.hadoop.hive.ql.Driver.compileInternal():1316
org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
org.apache.hive.service.cli.operation.SQLOperation.prepare():204
org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
org.apache.hive.service.cli.operation.Operation.run():320
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
sun.reflect.GeneratedMethodAccessor86.invoke():-1
sun.reflect.DelegatingMethodAccessorImpl.invoke():43
java.lang.reflect.Method.invoke():498
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1807
org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
com.sun.proxy.$Proxy42.executeStatementAsync():-1
org.apache.hive.service.cli.CLIService.executeStatementAsync():310
org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
org.apache.thrift.ProcessFunction.process():39
org.apache.thrift.TBaseProcessor.process():39
org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local>
{
"id": {
"part1": 2738948259170077000,
"part2": -61360705522985576
},
"type": 1,
"start": 1509772561834,
"end": 1509772562065,
"query": "SELECT count(*) FROM hive2.`contact`",
"plan": "00-00 Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01 Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02 Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
"foreman": {
"address": "ip-myip.mydomain.orghere.com",
"userPort": 31010,
"controlPort": 31011,
"dataPort": 31012,
"version": "1.11.0"
},
"state": 4,
"totalFragments": 1,
"finishedFragments": 0,
"fragmentProfile": [
{
"majorFragmentId": 0,
"minorFragmentProfile": [
{
"state": 0,
"minorFragmentId": 0,
"endpoint": {
"address": "ip-myip.mydomain.orghere.com",
"userPort": 31010,
"controlPort": 31011,
"dataPort": 31012,
"version": "1.11.0"
},
"lastUpdate": 1509772562054,
"lastProgress": 1509772562054
}
]
}
],
"user": "anonymous",
"error": "DATA_READ ERROR: Drill Remote Exception\n\n",
"verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
"errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
"errorNode": "ip-myip.mydomain.orghere.com:31010",
"optionsJson": "[ {\n \"kind\" : \"BOOLEAN\",\n \"type\" : \"SESSION\",\n \"name\" : \"exec.errors.verbose\",\n \"bool_val\" : true\n} ]",
"planEnd": 1509772562051,
"queueWaitEnd": 0
}
0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0
To prove that it can connect, here is the query plan, showing all columns from the table:
00-00 Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
00-01 Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
00-02 SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
00-03 Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
00-04 Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
00-05 Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
00-06 Jdbc(sql=[SELECT *
FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282
Below connection with beeline proves there is data in the table:
Beeline version 2.3.0 by Apache Hive
0: jdbc:hive2://localhost:10000/default> show create table contact;
+----------------------------------------------------+
| createtab_stmt |
+----------------------------------------------------+
| CREATE TABLE `contact`( |
| `contact_id` varchar(50), |
| `first_name` varchar(50), |
| `last_name` varchar(50), |
| `address_line_1` varchar(100), |
| `address_line_2` varchar(100), |
| `city` varchar(50), |
| `postcode` varchar(10), |
| `email` varchar(100), |
| `dob` date, |
| `gender` varchar(1), |
| `marital_status` varchar(1), |
| `tfn` varchar(20)) |
| ROW FORMAT SERDE |
| 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' |
| STORED AS INPUTFORMAT |
| 'org.apache.hadoop.mapred.TextInputFormat' |
| OUTPUTFORMAT |
| 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
| LOCATION |
| 'file:/home/ec2-user/warehouse/contact' |
| TBLPROPERTIES ( |
| 'transient_lastDdlTime'='1509487563') |
+----------------------------------------------------+
23 rows selected (0.361 seconds)
0: jdbc:hive2://localhost:10000/default> select * from contact;
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| contact.contact_id | contact.first_name | contact.last_name | contact.address_line_1 | contact.address_line_2 | contact.city | contact.postcode | contact.email | contact.dob | contact.gender | contact.marital_status | contact.tfn |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| C1 | Bob | B | St | | Ta | 2100 | bob@b.com | 2020-03-01 | M | M | 12 |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
1 rows selected (0.156 seconds)
0: jdbc:hive2://localhost:10000/default>
{code}
> HiveServer2 queries throw error with jdbc connection
> ----------------------------------------------------
>
> Key: DRILL-5932
> URL: https://issues.apache.org/jira/browse/DRILL-5932
> Project: Apache Drill
> Issue Type: Bug
> Components: Client - JDBC, Storage - Hive
> Affects Versions: 1.11.0
> Environment: linux
> 2.3 hive version
> Reporter: tooptoop4
> Priority: Blocker
>
> Basic hive queries all throw error!
> {code:sql}
> copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar
> added this storage plugin:
> {
> "type": "jdbc",
> "driver": "org.apache.hive.jdbc.HiveDriver",
> "url": "jdbc:hive2://host:10000/default",
> "username": "hive",
> "password": "hive1234",
> "enabled": true
> }
> [ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
> [ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
> OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
> Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
> INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
> apache drill 1.11.0
> "this isn't your grandfather's sql"
> 0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT COUNT(*) AS EXPR$0
> FROM (SELECT 0 AS $f0
> FROM.default.contact) AS t
> plugin hive2
> Fragment 0:0
> [Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)
> 0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
> +-------+-------------------------------+
> | ok | summary |
> +-------+-------------------------------+
> | true | exec.errors.verbose updated. |
> +-------+-------------------------------+
> 1 row selected (0.351 seconds)
> 0: jdbc:drill:zk=local> select * from hive2.contact;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT *
> FROM.default.contact
> plugin hive2
> Fragment 0:0
> [Error Id: fe36b026-e8ff-4354-af6c-6073130680c9 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]
> (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
> org.apache.hive.jdbc.Utils.verifySuccess():267
> org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
> org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
> org.apache.hive.jdbc.HiveStatement.execute():253
> org.apache.hive.jdbc.HiveStatement.executeQuery():476
> org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
> org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
> org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
> org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
> org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
> org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
> org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
> org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
> org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
> org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
> org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
> org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
> org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
> org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
> org.apache.drill.common.SelfCleaningRunnable.run():38
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
> org.apache.hive.service.cli.operation.Operation.toSQLException():380
> org.apache.hive.service.cli.operation.SQLOperation.prepare():206
> org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
> org.apache.hive.service.cli.operation.Operation.run():320
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
> sun.reflect.GeneratedMethodAccessor86.invoke():-1
> sun.reflect.DelegatingMethodAccessorImpl.invoke():43
> java.lang.reflect.Method.invoke():498
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1807
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
> com.sun.proxy.$Proxy42.executeStatementAsync():-1
> org.apache.hive.service.cli.CLIService.executeStatementAsync():310
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
> org.apache.thrift.ProcessFunction.process():39
> org.apache.thrift.TBaseProcessor.process():39
> org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
> org.apache.hadoop.hive.ql.Driver.compile():468
> org.apache.hadoop.hive.ql.Driver.compileInternal():1316
> org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
> org.apache.hive.service.cli.operation.SQLOperation.prepare():204
> org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
> org.apache.hive.service.cli.operation.Operation.run():320
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
> sun.reflect.GeneratedMethodAccessor86.invoke():-1
> sun.reflect.DelegatingMethodAccessorImpl.invoke():43
> java.lang.reflect.Method.invoke():498
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1807
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
> com.sun.proxy.$Proxy42.executeStatementAsync():-1
> org.apache.hive.service.cli.CLIService.executeStatementAsync():310
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
> org.apache.thrift.ProcessFunction.process():39
> org.apache.thrift.TBaseProcessor.process():39
> org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748 (state=,code=0)
> 0: jdbc:drill:zk=local> select * from hive2.`default`.contact;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT *
> FROM.default.contact
> plugin hive2
> Fragment 0:0
> [Error Id: 7b7aae92-fabd-49dd-ab84-15ca00189882 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]
> (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
> org.apache.hive.jdbc.Utils.verifySuccess():267
> org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
> org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
> org.apache.hive.jdbc.HiveStatement.execute():253
> org.apache.hive.jdbc.HiveStatement.executeQuery():476
> org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
> org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
> org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
> org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
> org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
> org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
> org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
> org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
> org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
> org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
> org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
> org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
> org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
> org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
> org.apache.drill.common.SelfCleaningRunnable.run():38
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
> org.apache.hive.service.cli.operation.Operation.toSQLException():380
> org.apache.hive.service.cli.operation.SQLOperation.prepare():206
> org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
> org.apache.hive.service.cli.operation.Operation.run():320
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
> sun.reflect.GeneratedMethodAccessor86.invoke():-1
> sun.reflect.DelegatingMethodAccessorImpl.invoke():43
> java.lang.reflect.Method.invoke():498
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1807
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
> com.sun.proxy.$Proxy42.executeStatementAsync():-1
> org.apache.hive.service.cli.CLIService.executeStatementAsync():310
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
> org.apache.thrift.ProcessFunction.process():39
> org.apache.thrift.TBaseProcessor.process():39
> org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
> org.apache.hadoop.hive.ql.Driver.compile():468
> org.apache.hadoop.hive.ql.Driver.compileInternal():1316
> org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
> org.apache.hive.service.cli.operation.SQLOperation.prepare():204
> org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
> org.apache.hive.service.cli.operation.Operation.run():320
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
> sun.reflect.GeneratedMethodAccessor86.invoke():-1
> sun.reflect.DelegatingMethodAccessorImpl.invoke():43
> java.lang.reflect.Method.invoke():498
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1807
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
> com.sun.proxy.$Proxy42.executeStatementAsync():-1
> org.apache.hive.service.cli.CLIService.executeStatementAsync():310
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
> org.apache.thrift.ProcessFunction.process():39
> org.apache.thrift.TBaseProcessor.process():39
> org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748 (state=,code=0)
> 0: jdbc:drill:zk=local>
> 0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT COUNT(*) AS EXPR$0
> FROM (SELECT 0 AS $f0
> FROM.default.contact) AS t
> plugin hive2
> Fragment 0:0
> [Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]
> (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
> org.apache.hive.jdbc.Utils.verifySuccess():267
> org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
> org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
> org.apache.hive.jdbc.HiveStatement.execute():253
> org.apache.hive.jdbc.HiveStatement.executeQuery():476
> org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
> org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
> org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
> org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
> org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
> org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
> org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
> org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
> org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
> org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
> org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
> org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
> org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
> org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
> org.apache.drill.common.SelfCleaningRunnable.run():38
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
> org.apache.hive.service.cli.operation.Operation.toSQLException():380
> org.apache.hive.service.cli.operation.SQLOperation.prepare():206
> org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
> org.apache.hive.service.cli.operation.Operation.run():320
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
> sun.reflect.GeneratedMethodAccessor86.invoke():-1
> sun.reflect.DelegatingMethodAccessorImpl.invoke():43
> java.lang.reflect.Method.invoke():498
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1807
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
> com.sun.proxy.$Proxy42.executeStatementAsync():-1
> org.apache.hive.service.cli.CLIService.executeStatementAsync():310
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
> org.apache.thrift.ProcessFunction.process():39
> org.apache.thrift.TBaseProcessor.process():39
> org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
> org.apache.hadoop.hive.ql.Driver.compile():468
> org.apache.hadoop.hive.ql.Driver.compileInternal():1316
> org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
> org.apache.hive.service.cli.operation.SQLOperation.prepare():204
> org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
> org.apache.hive.service.cli.operation.Operation.run():320
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
> sun.reflect.GeneratedMethodAccessor86.invoke():-1
> sun.reflect.DelegatingMethodAccessorImpl.invoke():43
> java.lang.reflect.Method.invoke():498
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1807
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
> com.sun.proxy.$Proxy42.executeStatementAsync():-1
> org.apache.hive.service.cli.CLIService.executeStatementAsync():310
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
> org.apache.thrift.ProcessFunction.process():39
> org.apache.thrift.TBaseProcessor.process():39
> org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748 (state=,code=0)
> 0: jdbc:drill:zk=local>
> {
> "id": {
> "part1": 2738948259170077000,
> "part2": -61360705522985576
> },
> "type": 1,
> "start": 1509772561834,
> "end": 1509772562065,
> "query": "SELECT count(*) FROM hive2.`contact`",
> "plan": "00-00 Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01 Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02 Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
> "foreman": {
> "address": "ip-myip.mydomain.orghere.com",
> "userPort": 31010,
> "controlPort": 31011,
> "dataPort": 31012,
> "version": "1.11.0"
> },
> "state": 4,
> "totalFragments": 1,
> "finishedFragments": 0,
> "fragmentProfile": [
> {
> "majorFragmentId": 0,
> "minorFragmentProfile": [
> {
> "state": 0,
> "minorFragmentId": 0,
> "endpoint": {
> "address": "ip-myip.mydomain.orghere.com",
> "userPort": 31010,
> "controlPort": 31011,
> "dataPort": 31012,
> "version": "1.11.0"
> },
> "lastUpdate": 1509772562054,
> "lastProgress": 1509772562054
> }
> ]
> }
> ],
> "user": "anonymous",
> "error": "DATA_READ ERROR: Drill Remote Exception\n\n",
> "verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
> "errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
> "errorNode": "ip-myip.mydomain.orghere.com:31010",
> "optionsJson": "[ {\n \"kind\" : \"BOOLEAN\",\n \"type\" : \"SESSION\",\n \"name\" : \"exec.errors.verbose\",\n \"bool_val\" : true\n} ]",
> "planEnd": 1509772562051,
> "queueWaitEnd": 0
> }
> 0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT *
> FROM.default.contact
> plugin hive2
> Fragment 0:0
> To prove that it can connect, here is the query plan, showing all columns from the table:
> 00-00 Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
> 00-01 Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
> 00-02 SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
> 00-03 Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
> 00-04 Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
> 00-05 Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
> 00-06 Jdbc(sql=[SELECT *
> FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282
> Below connection with beeline proves there is data in the table:
> Beeline version 2.3.0 by Apache Hive
> 0: jdbc:hive2://localhost:10000/default> show create table contact;
> +----------------------------------------------------+
> | createtab_stmt |
> +----------------------------------------------------+
> | CREATE TABLE `contact`( |
> | `contact_id` varchar(50), |
> | `first_name` varchar(50), |
> | `last_name` varchar(50), |
> | `address_line_1` varchar(100), |
> | `address_line_2` varchar(100), |
> | `city` varchar(50), |
> | `postcode` varchar(10), |
> | `email` varchar(100), |
> | `dob` date, |
> | `gender` varchar(1), |
> | `marital_status` varchar(1), |
> | `tfn` varchar(20)) |
> | ROW FORMAT SERDE |
> | 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' |
> | STORED AS INPUTFORMAT |
> | 'org.apache.hadoop.mapred.TextInputFormat' |
> | OUTPUTFORMAT |
> | 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
> | LOCATION |
> | 'file:/home/ec2-user/warehouse/contact' |
> | TBLPROPERTIES ( |
> | 'transient_lastDdlTime'='1509487563') |
> +----------------------------------------------------+
> 23 rows selected (0.361 seconds)
> 0: jdbc:hive2://localhost:10000/default> select * from contact;
> +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
> | contact.contact_id | contact.first_name | contact.last_name | contact.address_line_1 | contact.address_line_2 | contact.city | contact.postcode | contact.email | contact.dob | contact.gender | contact.marital_status | contact.tfn |
> +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
> | C1 | Bob | B | St | | Ta | 2100 | bob@b.com | 2020-03-01 | M | M | 12 |
> +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
> 1 rows selected (0.156 seconds)
> 0: jdbc:hive2://localhost:10000/default>
> {code}
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)