You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@drill.apache.org by "Aditya Kishore (JIRA)" <ji...@apache.org> on 2014/05/22 03:21:38 UTC

[jira] [Updated] (DRILL-802) convert_to(int_field, 'INT_HADOOPV') throws IndexOutOfBoundsException

     [ https://issues.apache.org/jira/browse/DRILL-802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Aditya Kishore updated DRILL-802:
---------------------------------

    Summary: convert_to(int_field, 'INT_HADOOPV') throws IndexOutOfBoundsException  (was: convert int to int_hadoopv throws indexoutofboundexception)

> convert_to(int_field, 'INT_HADOOPV') throws IndexOutOfBoundsException
> ---------------------------------------------------------------------
>
>                 Key: DRILL-802
>                 URL: https://issues.apache.org/jira/browse/DRILL-802
>             Project: Apache Drill
>          Issue Type: Bug
>          Components: Execution - Operators
>            Reporter: Chun Chang
>            Assignee: Aditya Kishore
>
> A query converting INT to INT_HADOOPV caused the exception:
> 0: jdbc:drill:schema=dfs> select c_row, c_int, convert_from(convert_to(c_int, 'INT_HADOOPV'), 'INT_HADOOPV') from data;
> Query failed: org.apache.drill.exec.rpc.RpcException: Remote failure while running query.[error_id: "2460e868-70a8-4133-9aa4-2884140cfb8b"
> endpoint {
>   address: "qa-node117.qa.lab"
>   user_port: 31010
>   control_port: 31011
>   data_port: 31012
> }
> error_type: 0
> message: "Failure while running fragment. < IndexOutOfBoundsException:[ writerIndex: 0 (expected: readerIndex(1) <= writerIndex <= capacity(9)) ]"
> ]
> Error: exception while executing query (state=,code=0)
> column c_int holds the following int value:
> 0: jdbc:drill:schema=dfs> select c_row, c_int from data;
> +------------+------------+
> |   c_row    |   c_int    |
> +------------+------------+
> | 1          | 0          |
> | 2          | 1          |
> | 3          | -1         |
> | 4          | 12         |
> | 5          | 123        |
> | 6          | 92032039   |
> | 7          | -23395000  |
> | 8          | -99392039  |
> | 9          | -2147483648 |
> | 10         | 2147483647 |
> | 11         | 32767      |
> | 12         | -32767     |
> | 13         | 49032      |
> | 14         | -4989385   |
> | 15         | 69834830   |
> | 16         | 243        |
> | 17         | -426432    |
> | 18         | -3904      |
> | 19         | 489392758  |
> | 20         | 589032574  |
> | 21         | 340000504  |
> | 22         | 0          |
> | 23         | 1          |
> +------------+------------+
> stack trace:
> 15:33:58.248 [c8d7eb9d-21af-4e1b-bc61-6f3349915d9a:frag:0:0] DEBUG o.a.d.e.w.fragment.FragmentExecutor - Caught exception while running fragment
> java.lang.IndexOutOfBoundsException: writerIndex: 0 (expected: readerIndex(1) <= writerIndex <= capacity(9))
>   at io.netty.buffer.AbstractByteBuf.writerIndex(AbstractByteBuf.java:87) ~[netty-buffer-4.0.7.Final.jar:na]
>   at io.netty.buffer.SwappedByteBuf.writerIndex(SwappedByteBuf.java:112) ~[netty-buffer-4.0.7.Final.jar:na]
>   at org.apache.drill.exec.util.ConvertUtil$HadoopWritables.writeVLong(ConvertUtil.java:96) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.test.generated.ProjectorGen722.doEval(ProjectorTemplate.java:40) ~[na:na]
>   at org.apache.drill.exec.test.generated.ProjectorGen722.projectRecords(ProjectorTemplate.java:65) ~[na:na]
>   at org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.doWork(ProjectRecordBatch.java:94) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.record.AbstractSingleRecordBatch.next(AbstractSingleRecordBatch.java:66) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.next(ProjectRecordBatch.java:83) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.physical.impl.validate.IteratorValidatorBatchIterator.next(IteratorValidatorBatchIterator.java:111) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.next(ScreenCreator.java:80) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.work.fragment.FragmentExecutor.run(FragmentExecutor.java:104) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [na:1.7.0_45]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [na:1.7.0_45]
>   at java.lang.Thread.run(Thread.java:744) [na:1.7.0_45]
> 15:33:58.249 [c8d7eb9d-21af-4e1b-bc61-6f3349915d9a:frag:0:0] ERROR o.a.d.e.w.f.AbstractStatusReporter - Error 2460e868-70a8-4133-9aa4-2884140cfb8b: Failure while running fragment.
> java.lang.IndexOutOfBoundsException: writerIndex: 0 (expected: readerIndex(1) <= writerIndex <= capacity(9))
>   at io.netty.buffer.AbstractByteBuf.writerIndex(AbstractByteBuf.java:87) ~[netty-buffer-4.0.7.Final.jar:na]
>   at io.netty.buffer.SwappedByteBuf.writerIndex(SwappedByteBuf.java:112) ~[netty-buffer-4.0.7.Final.jar:na]
>   at org.apache.drill.exec.util.ConvertUtil$HadoopWritables.writeVLong(ConvertUtil.java:96) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.test.generated.ProjectorGen722.doEval(ProjectorTemplate.java:40) ~[na:na]
>   at org.apache.drill.exec.test.generated.ProjectorGen722.projectRecords(ProjectorTemplate.java:65) ~[na:na]
>   at org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.doWork(ProjectRecordBatch.java:94) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.record.AbstractSingleRecordBatch.next(AbstractSingleRecordBatch.java:66) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.next(ProjectRecordBatch.java:83) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.physical.impl.validate.IteratorValidatorBatchIterator.next(IteratorValidatorBatchIterator.java:111) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.next(ScreenCreator.java:80) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at org.apache.drill.exec.work.fragment.FragmentExecutor.run(FragmentExecutor.java:104) ~[drill-java-exec-1.0.0-m2-incubating-SNAPSHOT-rebuffed.jar:1.0.0-m2-incubating-SNAPSHOT]
>   at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [na:1.7.0_45]
>   at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [na:1.7.0_45]
>   at java.lang.Thread.run(Thread.java:744) [na:1.7.0_45]
> physical plan:
> 0: jdbc:drill:schema=dfs> explain plan for select c_row, c_int, convert_from(convert_to(c_int, 'INT_HADOOPV'), 'INT_HADOOPV') from data;
> +------------+------------+
> |    text    |    json    |
> +------------+------------+
> | ScreenPrel
>   ProjectPrel(c_row=[$1], c_int=[$2], EXPR$2=[CONVERT_FROM(CONVERT_TO($2, 'INT_HADOOPV'), 'INT_HADOOPV')])
>     ScanPrel(groupscan=[ParquetGroupScan [entries=[ReadEntryWithPath [path=maprfs:/user/root/mondrian/data]], selectionRoot=/user/root/mondrian/data, columns=[SchemaPath [`c_row`], SchemaPath [`c_int`]]]])
>  | {
>   "head" : {
>     "version" : 1,
>     "generator" : {
>       "type" : "ExplainHandler",
>       "info" : ""
>     },
>     "type" : "APACHE_DRILL_PHYSICAL",
>     "options" : [ ],
>     "resultMode" : "EXEC"
>   },
>   "graph" : [ {
>     "pop" : "parquet-scan",
>     "@id" : 1,
>     "entries" : [ {
>       "path" : "maprfs:/user/root/mondrian/data"
>     } ],
>     "storage" : {
>       "type" : "file",
>       "connection" : "maprfs:///",
>       "workspaces" : {
>         "default" : {
>           "location" : "/user/root/mondrian/",
>           "writable" : false,
>           "storageformat" : null
>         },
>         "home" : {
>           "location" : "/",
>           "writable" : false,
>           "storageformat" : null
>         },
>         "root" : {
>           "location" : "/",
>           "writable" : false,
>           "storageformat" : null
>         },
>         "tmp" : {
>           "location" : "/tmp",
>           "writable" : true,
>           "storageformat" : "csv"
>         }
>       },
>       "formats" : {
>         "psv" : {
>           "type" : "text",
>           "extensions" : [ "tbl" ],
>           "delimiter" : "|"
>         },
>         "csv" : {
>           "type" : "text",
>           "extensions" : [ "csv" ],
>           "delimiter" : ","
>         },
>         "tsv" : {
>           "type" : "text",
>           "extensions" : [ "tsv" ],
>           "delimiter" : "\t"
>         },
>         "parquet" : {
>           "type" : "parquet"
>         },
>         "json" : {
>           "type" : "json"
>         }
>       }
>     },
>     "format" : {
>       "type" : "parquet"
>     },
>     "columns" : [ "`c_row`", "`c_int`" ],
>     "selectionRoot" : "/user/root/mondrian/data"
>   }, {
>     "pop" : "project",
>     "@id" : 2,
>     "exprs" : [ {
>       "ref" : "`c_row`",
>       "expr" : "`c_row`"
>     }, {
>       "ref" : "`c_int`",
>       "expr" : "`c_int`"
>     }, {
>       "ref" : "`EXPR$2`",
>       "expr" : "convert_from(convert_to(`c_int`, \"INT_HADOOPV\"), \"INT_HADOOPV\")"
>     } ],
>     "child" : 1,
>     "initialAllocation" : 1000000,
>     "maxAllocation" : 10000000000
>   }, {
>     "pop" : "screen",
>     "@id" : 3,
>     "child" : 2,
>     "initialAllocation" : 1000000,
>     "maxAllocation" : 10000000000
>   } ]
> } |



--
This message was sent by Atlassian JIRA
(v6.2#6252)