You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@carbondata.apache.org by "Babulal (JIRA)" <ji...@apache.org> on 2018/07/13 16:01:00 UTC

[jira] [Created] (CARBONDATA-2743) [MV] After MV creation limit queries Throws Exceptions for table which does not have mv data map

Babulal created CARBONDATA-2743:
-----------------------------------

             Summary: [MV] After MV creation limit queries Throws Exceptions for table which does not have mv data map
                 Key: CARBONDATA-2743
                 URL: https://issues.apache.org/jira/browse/CARBONDATA-2743
             Project: CarbonData
          Issue Type: Bug
            Reporter: Babulal


0: jdbc:hive2://10.18.16.173:23040/default> create table mytest_50_s13 (name string,rownumber string, m1 float) stored by 'carbondata' TBLPROPERTIES('sort_scope'='global_sort');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (12.209 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> load data inpath 'hdfs://hacluster/tmp/data/cbo_1.csv' into table mytest_50_s13 options('FILEHEADER'='name,rownumber,m1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (79.901 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> create datamap map10 using 'mv' as select sum(m1),rownumber from mytest_50_s13 group by rownumber;
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (17.05 seconds)

0: jdbc:hive2://10.18.16.173:23040/default> show datamap on table mytest_50_s13;
+--------------+------------+-------------------+---------------------+--+
| DataMapName  | ClassName  | Associated Table  | DataMap Properties  |
+--------------+------------+-------------------+---------------------+--+
| map10        | mv         | babu.map10_table  |                     |
+--------------+------------+-------------------+---------------------+--+
1 row selected (0.815 seconds)



Now create one more table without mv

: jdbc:hive2://10.18.16.173:23040/default> create table mytest_50_s14 (name string,rownumber string, m1 float) stored by 'carbondata' TBLPROPERTIES('sort_scope'='global_sort');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (12.209 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> load data inpath 'hdfs://hacluster/tmp/data/cbo_1.csv'' into table mytest_50_s14 options('FILEHEADER'='name,rownumber,m1');
+---------+--+
| Result  |
+---------+--+
+---------+--+
No rows selected (79.901 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> select * from mytest_50_s14 limit 10;
Error: java.lang.UnsupportedOperationException: unsupported operation: Modular plan not supported (e.g. has subquery expression) for
GlobalLimit 10
+- LocalLimit 10
   +- Relation[name#1026,rownumber#1027,m1#1028] CarbonDatasourceHadoopRelation [ Database name :babu, Table name :mytest_50_s14, Schema :Some(StructType(StructField(name,StringType,true), StructField(rownumber,StringType,true), StructField(m1,DoubleType,true))) ] (state=,code=0)





2018-07-13 00:42:51,540 | INFO  | [pool-25-thread-32] | OperationId=b5c2c8b2-1ef4-4894-a709-2a738bd81f76        Result=FAIL | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:280)
2018-07-13 00:42:51,540 | ERROR | [pool-25-thread-32] | Error executing query, currentState RUNNING,  | org.apache.spark.internal.Logging$class.logError(Logging.scala:91)
java.lang.UnsupportedOperationException: unsupported operation: Modular plan not supported (e.g. has subquery expression) for
GlobalLimit 10
+- LocalLimit 10
   +- Relation[name#1026,rownumber#1027,m1#1028] CarbonDatasourceHadoopRelation [ Database name :babu, Table name :mytest_50_s14, Schema :Some(StructType(StructField(name,StringType,true), StructField(rownumber,StringType,true), StructField(m1,DoubleType,true))) ]

        at org.apache.carbondata.mv.plans.package$.supports(package.scala:52)
        at org.apache.carbondata.mv.plans.modular.Modularizer.org$apache$carbondata$mv$plans$modular$Modularizer$$modularizeCore(Modularizer.scala:102)
        at org.apache.carbondata.mv.plans.modular.Modularizer.modularize(Modularizer.scala:65)
        at org.apache.carbondata.mv.rewrite.QueryRewrite.modularPlan$lzycompute(QueryRewrite.scala:50)
        at org.apache.carbondata.mv.rewrite.QueryRewrite.modularPlan(QueryRewrite.scala:49)
        at org.apache.carbondata.mv.rewrite.QueryRewrite.withSummaryData$lzycompute(QueryRewrite.scala:53)
        at org.apache.carbondata.mv.rewrite.QueryRewrite.withSummaryData(QueryRewrite.scala:52)
        at org.apache.carbondata.mv.rewrite.QueryRewrite.withMVTable$lzycompute(QueryRewrite.scala:55)
        at org.apache.carbondata.mv.rewrite.QueryRewrite.withMVTable(QueryRewrite.scala:55)
        at org.apache.carbondata.mv.datamap.MVAnalyzerRule.apply(MVAnalyzerRule.scala:68)
        at org.apache.carbondata.mv.datamap.MVAnalyzerRule.apply(MVAnalyzerRule.scala:38)
        at org.apache.spark.sql.hive.CarbonAnalyzer.execute(CarbonAnalyzer.scala:46)
        at org.apache.spark.sql.hive.CarbonAnalyzer.execute(CarbonAnalyzer.scala:27)
        at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:75)
        at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:73)
        at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:56)
        at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67)
        at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:632)
        at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:691)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:244)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:176)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:173)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1778)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:186)
        at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
2018-07-13 00:42:51,541 | ERROR | [pool-25-thread-32] | Error running hive query:  | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:180)
org.apache.hive.service.cli.HiveSQLException: java.lang.UnsupportedOperationException: unsupported operation: Modular plan not supported (e.g. has subquery expression) for
GlobalLimit 10
+- LocalLimit 10
   +- Relation[name#1026,rownumber#1027,m1#1028] CarbonDatasourceHadoopRelation [ Database name :babu, Table name :mytest_50_s14, Schema :Some(StructType(StructField(name,StringType,true), StructField(rownumber,StringType,true), StructField(m1,DoubleType,true))) ]

        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:289)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:176)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:173)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1778)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStat



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)