Uploaded image for project: 'Apache Drill'
  1. Apache Drill
  2. DRILL-5932

HiveServer2 queries throw error with jdbc connection

    XMLWordPrintableJSON

    Details

    • Type: Bug
    • Status: Open
    • Priority: Blocker
    • Resolution: Unresolved
    • Affects Version/s: 1.11.0
    • Fix Version/s: None
    • Labels:
      None
    • Environment:

      linux
      2.3 hive version

    • Flags:
      Important

      Description

      Basic hive queries all throw error!

      
      copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar
      
      added this storage plugin:
      
      {
        "type": "jdbc",
        "driver": "org.apache.hive.jdbc.HiveDriver",
        "url": "jdbc:hive2://host:10000/default",
        "username": "hive",
        "password": "hive1234",
        "enabled": true
      }
      
      [ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
      [ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
      OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
      Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
      INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
      apache drill 1.11.0
      "this isn't your grandfather's sql"
      0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
      Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
      
      sql SELECT COUNT(*) AS EXPR$0
      FROM (SELECT 0 AS $f0
      FROM.default.contact) AS t
      plugin hive2
      Fragment 0:0
      
      [Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)
      
      
      
      
      0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
      +-------+-------------------------------+
      |  ok   |            summary            |
      +-------+-------------------------------+
      | true  | exec.errors.verbose updated.  |
      +-------+-------------------------------+
      1 row selected (0.351 seconds)
      
      0: jdbc:drill:zk=local> select * from hive2.contact;
      Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
      
      sql SELECT *
      FROM.default.contact
      plugin hive2
      Fragment 0:0
      
      [Error Id: fe36b026-e8ff-4354-af6c-6073130680c9 on ip-ip.domain.org.com:31010]
      
        (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
          org.apache.hive.jdbc.Utils.verifySuccess():267
          org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
          org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
          org.apache.hive.jdbc.HiveStatement.execute():253
          org.apache.hive.jdbc.HiveStatement.executeQuery():476
          org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
          org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
          org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
          org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
          org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
          org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
          org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
          org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
          org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
          org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
          org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
          org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
          org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
          org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
          org.apache.drill.common.SelfCleaningRunnable.run():38
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748
        Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
          org.apache.hive.service.cli.operation.Operation.toSQLException():380
          org.apache.hive.service.cli.operation.SQLOperation.prepare():206
          org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
          org.apache.hive.service.cli.operation.Operation.run():320
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
          sun.reflect.GeneratedMethodAccessor86.invoke():-1
          sun.reflect.DelegatingMethodAccessorImpl.invoke():43
          java.lang.reflect.Method.invoke():498
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
          org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
          org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
          java.security.AccessController.doPrivileged():-2
          javax.security.auth.Subject.doAs():422
          org.apache.hadoop.security.UserGroupInformation.doAs():1807
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
          com.sun.proxy.$Proxy42.executeStatementAsync():-1
          org.apache.hive.service.cli.CLIService.executeStatementAsync():310
          org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
          org.apache.thrift.ProcessFunction.process():39
          org.apache.thrift.TBaseProcessor.process():39
          org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
          org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748
        Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
          org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
          org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
          org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
          org.apache.hadoop.hive.ql.Driver.compile():468
          org.apache.hadoop.hive.ql.Driver.compileInternal():1316
          org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
          org.apache.hive.service.cli.operation.SQLOperation.prepare():204
          org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
          org.apache.hive.service.cli.operation.Operation.run():320
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
          sun.reflect.GeneratedMethodAccessor86.invoke():-1
          sun.reflect.DelegatingMethodAccessorImpl.invoke():43
          java.lang.reflect.Method.invoke():498
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
          org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
          org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
          java.security.AccessController.doPrivileged():-2
          javax.security.auth.Subject.doAs():422
          org.apache.hadoop.security.UserGroupInformation.doAs():1807
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
          com.sun.proxy.$Proxy42.executeStatementAsync():-1
          org.apache.hive.service.cli.CLIService.executeStatementAsync():310
          org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
          org.apache.thrift.ProcessFunction.process():39
          org.apache.thrift.TBaseProcessor.process():39
          org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
          org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748 (state=,code=0)
      0: jdbc:drill:zk=local> select * from hive2.`default`.contact;
      Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
      
      sql SELECT *
      FROM.default.contact
      plugin hive2
      Fragment 0:0
      
      [Error Id: 7b7aae92-fabd-49dd-ab84-15ca00189882 on ip-ip.domain.org.com:31010]
      
        (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
          org.apache.hive.jdbc.Utils.verifySuccess():267
          org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
          org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
          org.apache.hive.jdbc.HiveStatement.execute():253
          org.apache.hive.jdbc.HiveStatement.executeQuery():476
          org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
          org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
          org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
          org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
          org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
          org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
          org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
          org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
          org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
          org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
          org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
          org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
          org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
          org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
          org.apache.drill.common.SelfCleaningRunnable.run():38
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748
        Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
          org.apache.hive.service.cli.operation.Operation.toSQLException():380
          org.apache.hive.service.cli.operation.SQLOperation.prepare():206
          org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
          org.apache.hive.service.cli.operation.Operation.run():320
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
          sun.reflect.GeneratedMethodAccessor86.invoke():-1
          sun.reflect.DelegatingMethodAccessorImpl.invoke():43
          java.lang.reflect.Method.invoke():498
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
          org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
          org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
          java.security.AccessController.doPrivileged():-2
          javax.security.auth.Subject.doAs():422
          org.apache.hadoop.security.UserGroupInformation.doAs():1807
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
          com.sun.proxy.$Proxy42.executeStatementAsync():-1
          org.apache.hive.service.cli.CLIService.executeStatementAsync():310
          org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
          org.apache.thrift.ProcessFunction.process():39
          org.apache.thrift.TBaseProcessor.process():39
          org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
          org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748
        Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
          org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
          org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
          org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
          org.apache.hadoop.hive.ql.Driver.compile():468
          org.apache.hadoop.hive.ql.Driver.compileInternal():1316
          org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
          org.apache.hive.service.cli.operation.SQLOperation.prepare():204
          org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
          org.apache.hive.service.cli.operation.Operation.run():320
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
          sun.reflect.GeneratedMethodAccessor86.invoke():-1
          sun.reflect.DelegatingMethodAccessorImpl.invoke():43
          java.lang.reflect.Method.invoke():498
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
          org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
          org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
          java.security.AccessController.doPrivileged():-2
          javax.security.auth.Subject.doAs():422
          org.apache.hadoop.security.UserGroupInformation.doAs():1807
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
          com.sun.proxy.$Proxy42.executeStatementAsync():-1
          org.apache.hive.service.cli.CLIService.executeStatementAsync():310
          org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
          org.apache.thrift.ProcessFunction.process():39
          org.apache.thrift.TBaseProcessor.process():39
          org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
          org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748 (state=,code=0)
      0: jdbc:drill:zk=local>
      
      
      0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
      Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
      
      sql SELECT COUNT(*) AS EXPR$0
      FROM (SELECT 0 AS $f0
      FROM.default.contact) AS t
      plugin hive2
      Fragment 0:0
      
      [Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]
      
        (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
          org.apache.hive.jdbc.Utils.verifySuccess():267
          org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
          org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
          org.apache.hive.jdbc.HiveStatement.execute():253
          org.apache.hive.jdbc.HiveStatement.executeQuery():476
          org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
          org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
          org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
          org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
          org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
          org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
          org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
          org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
          org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
          org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
          org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
          org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
          org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
          org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
          org.apache.drill.common.SelfCleaningRunnable.run():38
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748
        Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
          org.apache.hive.service.cli.operation.Operation.toSQLException():380
          org.apache.hive.service.cli.operation.SQLOperation.prepare():206
          org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
          org.apache.hive.service.cli.operation.Operation.run():320
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
          sun.reflect.GeneratedMethodAccessor86.invoke():-1
          sun.reflect.DelegatingMethodAccessorImpl.invoke():43
          java.lang.reflect.Method.invoke():498
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
          org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
          org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
          java.security.AccessController.doPrivileged():-2
          javax.security.auth.Subject.doAs():422
          org.apache.hadoop.security.UserGroupInformation.doAs():1807
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
          com.sun.proxy.$Proxy42.executeStatementAsync():-1
          org.apache.hive.service.cli.CLIService.executeStatementAsync():310
          org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
          org.apache.thrift.ProcessFunction.process():39
          org.apache.thrift.TBaseProcessor.process():39
          org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
          org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748
        Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
          org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
          org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
          org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
          org.apache.hadoop.hive.ql.Driver.compile():468
          org.apache.hadoop.hive.ql.Driver.compileInternal():1316
          org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
          org.apache.hive.service.cli.operation.SQLOperation.prepare():204
          org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
          org.apache.hive.service.cli.operation.Operation.run():320
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
          org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
          sun.reflect.GeneratedMethodAccessor86.invoke():-1
          sun.reflect.DelegatingMethodAccessorImpl.invoke():43
          java.lang.reflect.Method.invoke():498
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
          org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
          org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
          java.security.AccessController.doPrivileged():-2
          javax.security.auth.Subject.doAs():422
          org.apache.hadoop.security.UserGroupInformation.doAs():1807
          org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
          com.sun.proxy.$Proxy42.executeStatementAsync():-1
          org.apache.hive.service.cli.CLIService.executeStatementAsync():310
          org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
          org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
          org.apache.thrift.ProcessFunction.process():39
          org.apache.thrift.TBaseProcessor.process():39
          org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
          org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
          java.util.concurrent.ThreadPoolExecutor.runWorker():1149
          java.util.concurrent.ThreadPoolExecutor$Worker.run():624
          java.lang.Thread.run():748 (state=,code=0)
      0: jdbc:drill:zk=local>
      
      
      {
          "id": {
              "part1": 2738948259170077000,
              "part2": -61360705522985576
          },
          "type": 1,
          "start": 1509772561834,
          "end": 1509772562065,
          "query": "SELECT count(*) FROM hive2.`contact`",
          "plan": "00-00    Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01      Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02        Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
          "foreman": {
              "address": "ip-myip.mydomain.orghere.com",
              "userPort": 31010,
              "controlPort": 31011,
              "dataPort": 31012,
              "version": "1.11.0"
          },
          "state": 4,
          "totalFragments": 1,
          "finishedFragments": 0,
          "fragmentProfile": [
              {
                  "majorFragmentId": 0,
                  "minorFragmentProfile": [
                      {
                          "state": 0,
                          "minorFragmentId": 0,
                          "endpoint": {
                              "address": "ip-myip.mydomain.orghere.com",
                              "userPort": 31010,
                              "controlPort": 31011,
                              "dataPort": 31012,
                              "version": "1.11.0"
                          },
                          "lastUpdate": 1509772562054,
                          "lastProgress": 1509772562054
                      }
                  ]
              }
          ],
          "user": "anonymous",
          "error": "DATA_READ ERROR: Drill Remote Exception\n\n",
          "verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
          "errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
          "errorNode": "ip-myip.mydomain.orghere.com:31010",
          "optionsJson": "[ {\n  \"kind\" : \"BOOLEAN\",\n  \"type\" : \"SESSION\",\n  \"name\" : \"exec.errors.verbose\",\n  \"bool_val\" : true\n} ]",
          "planEnd": 1509772562051,
          "queueWaitEnd": 0
      }
      
      
      
      0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
      Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
      
      sql SELECT *
      FROM.default.contact
      plugin hive2
      Fragment 0:0
      
      
      To prove that it can connect, here is the query plan, showing all columns from the table:
      
      
      00-00    Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
      00-01      Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
      00-02        SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
      00-03          Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
      00-04            Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
      00-05              Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
      00-06                Jdbc(sql=[SELECT *
      FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282
      
      
      Below connection with beeline proves there is data in the table:
      
      Beeline version 2.3.0 by Apache Hive
      0: jdbc:hive2://localhost:10000/default> show create table contact;
      +----------------------------------------------------+
      |                   createtab_stmt                   |
      +----------------------------------------------------+
      | CREATE TABLE `contact`(                            |
      |   `contact_id` varchar(50),                        |
      |   `first_name` varchar(50),                        |
      |   `last_name` varchar(50),                         |
      |   `address_line_1` varchar(100),                   |
      |   `address_line_2` varchar(100),                   |
      |   `city` varchar(50),                              |
      |   `postcode` varchar(10),                          |
      |   `email` varchar(100),                            |
      |   `dob` date,                                      |
      |   `gender` varchar(1),                             |
      |   `marital_status` varchar(1),                     |
      |   `tfn` varchar(20))                               |
      | ROW FORMAT SERDE                                   |
      |   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'  |
      | STORED AS INPUTFORMAT                              |
      |   'org.apache.hadoop.mapred.TextInputFormat'       |
      | OUTPUTFORMAT                                       |
      |   'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
      | LOCATION                                           |
      |   'file:/home/ec2-user/warehouse/contact'          |
      | TBLPROPERTIES (                                    |
      |   'transient_lastDdlTime'='1509487563')            |
      +----------------------------------------------------+
      23 rows selected (0.361 seconds)
      0: jdbc:hive2://localhost:10000/default> select * from contact;
      +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
      | contact.contact_id  | contact.first_name  | contact.last_name  | contact.address_line_1  | contact.address_line_2  | contact.city  | contact.postcode  |   contact.email    | contact.dob  | contact.gender  | contact.marital_status  | contact.tfn  |
      +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
      | C1                  | Bob                 | B                  | St                      |                         | Ta            | 2100              | bob@b.com          | 2020-03-01   | M               | M                       | 12           |
      +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
      1 rows selected (0.156 seconds)
      0: jdbc:hive2://localhost:10000/default>
      
      

        Attachments

          Activity

            People

            • Assignee:
              Unassigned
              Reporter:
              tooptoop4 tooptoop4
            • Votes:
              1 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

              • Created:
                Updated: