You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@carbondata.apache.org by "Vandana Yadav (JIRA)" <ji...@apache.org> on 2018/02/27 11:37:00 UTC

[jira] [Created] (CARBONDATA-2210) Not able to rename a partitioned table on cluster

Vandana Yadav created CARBONDATA-2210:
-----------------------------------------

             Summary: Not able to rename a partitioned table on cluster
                 Key: CARBONDATA-2210
                 URL: https://issues.apache.org/jira/browse/CARBONDATA-2210
             Project: CarbonData
          Issue Type: Bug
          Components: data-query
    Affects Versions: 1.4.0
         Environment: spark 2.2.1
            Reporter: Vandana Yadav


Not able to rename a partitioned table on cluster

Steps to reproduce:

1) Create a hive table:

  CREATE TABLE uniqdata_hive (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 int)ROW FORMAT DELIMITED FIELDS TERMINATED BY ','

LOAD DATA LOCAL INPATH '/opt/Carbon/CarbonData/TestData/Data/uniqdata/2000_UniqData.csv' into table UNIQDATA_HIVE

2) Create a carbon table:

CREATE TABLE uniqdata_int (CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 int) Partitioned by (cust_id int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB")

insert into uniqdata_int partition(cust_id)select * from uniqdata_hive limit 2

3) Execute Query:

alter table uniqdata_int rename to uniqdata_bigint1

Expected Result: table should be renamed successfully.

Actual Result:

operation failed for default.uniqdata_int: Alter table rename table operation failed: Folder rename failed for table default.uniqdata_int

 

logs 
[exec] 18/02/27 16:38:16 INFO SelectQuery: Executing Query: alter table uniqdata_int rename to uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO CarbonSparkSqlParser: Parsing command: alter table uniqdata_int rename to uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO CarbonLateDecodeRule: main skip CarbonOptimizer
     [exec] 18/02/27 16:38:16 INFO CarbonLateDecodeRule: main Skip CarbonOptimizer
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: int
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: array<string>
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_bigint1	
     [exec] 18/02/27 16:38:16 AUDIT CarbonAlterTableRenameCommand: [hadoop-master][root][Thread-1]Rename table request has been received for default.uniqdata_int
     [exec] 18/02/27 16:38:16 INFO CarbonAlterTableRenameCommand: main Rename table request has been received for default.uniqdata_int
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: int
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: array<string>
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: int
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: array<string>
     [exec] 18/02/27 16:38:16 INFO HdfsFileLock: main HDFS lock path:hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_int/meta.lock
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Trying to acquire lock: meta.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Successfully acquired the lock meta.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO HdfsFileLock: main HDFS lock path:hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_int/compaction.lock
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Trying to acquire lock: compaction.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Successfully acquired the lock compaction.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO HdfsFileLock: main HDFS lock path:hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_int/delete_segment.lock
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Trying to acquire lock: delete_segment.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Successfully acquired the lock delete_segment.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO HdfsFileLock: main HDFS lock path:hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_int/clean_files.lock
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Trying to acquire lock: clean_files.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Successfully acquired the lock clean_files.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO HdfsFileLock: main HDFS lock path:hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_int/droptable.lock
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Trying to acquire lock: droptable.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO CarbonLockUtil: main Successfully acquired the lock droptable.lockfor table: default_uniqdata_int
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: int
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: array<string>
     [exec] 18/02/27 16:38:16 INFO CarbonLRUCache: main Removed entry from InMemory lru cache :: default/uniqdata_int_/0/100000100001_batchno0-0-1519729675075.carbonindex
     [exec] 18/02/27 16:38:16 INFO CarbonLRUCache: main Removed entry from InMemory lru cache :: default/uniqdata_int_/0/100000100002_batchno0-0-1519729675075.carbonindex
     [exec] 18/02/27 16:38:16 INFO CarbonSparkSqlParser: Parsing command: `default`.`uniqdata_int`
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_database: default
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_database: default	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: int
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: array<string>
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_database: default
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_database: default	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: int
     [exec] 18/02/27 16:38:16 INFO CatalystSqlParser: Parsing command: array<string>
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_database: default
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_database: default	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_database: default
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_database: default	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_tables: db=default pat=*
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_tables: db=default pat=*	
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=Driver.run from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=compile from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=parse from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO ParseDriver: Parsing command: ALTER TABLE default.uniqdata_int RENAME TO default.uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO ParseDriver: Parse Completed
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=parse start=1519729696538 end=1519729696539 duration=1 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO Driver: Semantic Analysis Completed
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=semanticAnalyze start=1519729696539 end=1519729696542 duration=3 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Returning Hive schema: Schema(fieldSchemas:null, properties:null)
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=compile start=1519729696538 end=1519729696542 duration=4 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Concurrency mode is disabled, not creating a lock manager
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Starting command(queryId=root_20180227163816_c6a748b9-74de-4876-8686-13c92269add0): ALTER TABLE default.uniqdata_int RENAME TO default.uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=TimeToSubmit start=1519729696538 end=1519729696542 duration=4 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=runTasks from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Starting task [Stage-0:DDL] in serial mode
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_int
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_int	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: alter_table: db=default tbl=uniqdata_int newtbl=uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=alter_table: db=default tbl=uniqdata_int newtbl=uniqdata_bigint1	
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=runTasks start=1519729696542 end=1519729696568 duration=26 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=Driver.execute start=1519729696542 end=1519729696568 duration=26 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: OK
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=releaseLocks start=1519729696569 end=1519729696569 duration=0 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=Driver.run start=1519729696538 end=1519729696569 duration=31 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=releaseLocks start=1519729696569 end=1519729696569 duration=0 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=Driver.run from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=compile from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=parse from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO ParseDriver: Parsing command: ALTER TABLE default.uniqdata_bigint1 SET SERDEPROPERTIES('tableName'='uniqdata_bigint1', 'dbName'='default', 'tablePath'='hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_bigint1')
     [exec] 18/02/27 16:38:16 INFO ParseDriver: Parse Completed
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=parse start=1519729696569 end=1519729696570 duration=1 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_bigint1	
     [exec] 18/02/27 16:38:16 INFO Driver: Semantic Analysis Completed
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=semanticAnalyze start=1519729696570 end=1519729696573 duration=3 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Returning Hive schema: Schema(fieldSchemas:null, properties:null)
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=compile start=1519729696569 end=1519729696573 duration=4 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Concurrency mode is disabled, not creating a lock manager
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Starting command(queryId=root_20180227163816_4175ddb2-e8a0-4538-bab4-5688fe7b2a10): ALTER TABLE default.uniqdata_bigint1 SET SERDEPROPERTIES('tableName'='uniqdata_bigint1', 'dbName'='default', 'tablePath'='hdfs://hadoop-master:54311/opt/CarbonStore/default/uniqdata_bigint1')
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=TimeToSubmit start=1519729696569 end=1519729696573 duration=4 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=runTasks from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: Starting task [Stage-0:DDL] in serial mode
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=get_table : db=default tbl=uniqdata_bigint1	
     [exec] 18/02/27 16:38:16 INFO HiveMetaStore: 0: alter_table: db=default tbl=uniqdata_bigint1 newtbl=uniqdata_bigint1
     [exec] 18/02/27 16:38:16 INFO audit: ugi=root	ip=unknown-ip-addr	cmd=alter_table: db=default tbl=uniqdata_bigint1 newtbl=uniqdata_bigint1	
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=runTasks start=1519729696573 end=1519729696593 duration=20 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=Driver.execute start=1519729696573 end=1519729696593 duration=20 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO Driver: OK
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=releaseLocks start=1519729696594 end=1519729696594 duration=0 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=Driver.run start=1519729696569 end=1519729696594 duration=25 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 INFO PerfLogger: </PERFLOG method=releaseLocks start=1519729696594 end=1519729696594 duration=0 from=org.apache.hadoop.hive.ql.Driver>
     [exec] 18/02/27 16:38:16 ERROR HDFSCarbonFile: main Exception occured: rename destination directory is not empty: /opt/CarbonStore/default/uniqdata_bigint1
     [exec] 	at org.apache.hadoop.hdfs.server.namenode.FSDirRenameOp.validateOverwrite(FSDirRenameOp.java:529)
     [exec] 	at org.apache.hadoop.hdfs.server.namenode.FSDirRenameOp.unprotectedRenameTo(FSDirRenameOp.java:364)
     [exec] 	at org.apache.hadoop.hdfs.server.namenode.FSDirRenameOp.renameTo(FSDirRenameOp.java:282)
     [exec] 	at org.apache.hadoop.hdfs.server.namenode.FSDirRenameOp.renameToInt(FSDirRenameOp.java:247)
     [exec] 	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.renameTo(FSNamesystem.java:3675)
     [exec] 	at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.rename2(NameNodeRpcServer.java:913)
     [exec] 	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.rename2(ClientNamenodeProtocolServerSideTranslatorPB.java:587)
     [exec] 	at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
     [exec] 	at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
     [exec] 	at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
     [exec] 	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
     [exec] 	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
     [exec] 	at java.security.AccessController.doPrivileged(Native Method)
     [exec] 	at javax.security.auth.Subject.doAs(Subject.java:422)
     [exec] 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
     [exec] 	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
     [exec] 
     [exec] 18/02/27 16:38:16 ERROR CarbonAlterTableRenameCommand: main Rename table failed: Folder rename failed for table default.uniqdata_int
     [exec] java.lang.RuntimeException: Folder rename failed for table default.uniqdata_int
     [exec] 	at scala.sys.package$.error(package.scala:27)
     [exec] 	at org.apache.spark.sql.execution.command.schema.CarbonAlterTableRenameCommand.processMetadata(CarbonAlterTableRenameCommand.scala:138)
     [exec] 	at org.apache.spark.sql.execution.command.MetadataCommand.run(package.scala:68)
     [exec] 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
     [exec] 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
     [exec] 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:67)
     [exec] 	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:183)
     [exec] 	at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:68)
     [exec] 	at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:632)
     [exec] 	at com.huawei.spark.SessionManager.sql(SessionManager.java:42)
     [exec] 	at com.huawei.querymanagement.QueryManagement.sql(QueryManagement.java:62)
     [exec] 	at com.huawei.querymanagement.SelectQuery.testQuery(SelectQuery.java:70)
     [exec] 	at sun.reflect.GeneratedMethodAccessor88.invoke(Unknown Source)
     [exec] 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     [exec] 	at java.lang.reflect.Method.invoke(Method.java:498)
     [exec] 	at org.junit.internal.runners.TestMethod.invoke(TestMethod.java:59)
     [exec] 	at org.junit.internal.runners.MethodRoadie.runTestMethod(MethodRoadie.java:98)
     [exec] 	at org.junit.internal.runners.MethodRoadie$2.run(MethodRoadie.java:79)
     [exec] 	at org.junit.internal.runners.MethodRoadie.runBeforesThenTestThenAfters(MethodRoadie.java:87)
     [exec] 	at org.junit.internal.runners.MethodRoadie.runTest(MethodRoadie.java:77)
     [exec] 	at org.junit.internal.runners.MethodRoadie.run(MethodRoadie.java:42)
     [exec] 	at org.junit.internal.runners.JUnit4ClassRunner.invokeTestMethod(JUnit4ClassRunner.java:88)
     [exec] 	at org.junit.internal.runners.JUnit4ClassRunner.runMethods(JUnit4ClassRunner.java:51)
     [exec] 	at org.junit.runners.Parameterized$TestClassRunnerForParameters.run(Parameterized.java:98)
     [exec] 	at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
     [exec] 	at org.junit.runners.Parameterized.access$000(Parameterized.java:55)
     [exec] 	at org.junit.runners.Parameterized$1.run(Parameterized.java:131)
     [exec] 	at org.junit.internal.runners.ClassRoadie.runUnprotected(ClassRoadie.java:27)
     [exec] 	at org.junit.internal.runners.ClassRoadie.runProtected(ClassRoadie.java:37)
     [exec] 	at org.junit.runners.Parameterized.run(Parameterized.java:129)
     [exec] 	at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
     [exec] 	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:28)
     [exec] 	at org.junit.runner.JUnitCore.run(JUnitCore.java:130)
     [exec] 	at org.junit.runner.JUnitCore.run(JUnitCore.java:109)
     [exec] 	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
     [exec] 	at org.junit.runner.JUnitCore.runClasses(JUnitCore.java:60)
     [exec] 	at com.huawei.querymanagement.SelectQuerySuite.main(SelectQuerySuite.java:18)
     [exec] 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     [exec] 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
     [exec] 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     [exec] 	at java.lang.reflect.Method.invoke(Method.java:498)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:775)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
     [exec] 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
     [exec] 18/02/27 16:38:16 INFO AlterTableUtil$: main Alter table lock released successfully: meta.lock
     [exec] 18/02/27 16:38:16 INFO AlterTableUtil$: main Alter table lock released successfully: compaction.lock
     [exec] 18/02/27 16:38:16 INFO AlterTableUtil$: main Alter table lock released successfully: delete_segment.lock
     [exec] 18/02/27 16:38:16 INFO AlterTableUtil$: main Alter table lock released successfully: clean_files.lock
     [exec] 18/02/27 16:38:16 INFO AlterTableUtil$: main Alter table lock released successfully: droptable.lock
     [exec] 18/02/27 16:38:16 ERROR SelectQuery: An exception has occurred: 
     [exec] org.apache.carbondata.spark.exception.ProcessMetaDataException: operation failed for default.uniqdata_int: Alter table rename table operation failed: Folder rename failed for table default.uniqdata_int
     [exec] 	at org.apache.spark.sql.execution.command.MetadataProcessOpeation$class.throwMetadataException(package.scala:52)
     [exec] 	at org.apache.spark.sql.execution.command.MetadataCommand.throwMetadataException(package.scala:66)
     [exec] 	at org.apache.spark.sql.execution.command.schema.CarbonAlterTableRenameCommand.processMetadata(CarbonAlterTableRenameCommand.scala:174)
     [exec] 	at org.apache.spark.sql.execution.command.MetadataCommand.run(package.scala:68)
     [exec] 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
     [exec] 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
     [exec] 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:67)
     [exec] 	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:183)
     [exec] 	at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:68)
     [exec] 	at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:632)
     [exec] 	at com.huawei.spark.SessionManager.sql(SessionManager.java:42)
     [exec] 	at com.huawei.querymanagement.QueryManagement.sql(QueryManagement.java:62)
     [exec] 	at com.huawei.querymanagement.SelectQuery.testQuery(SelectQuery.java:70)
     [exec] 	at sun.reflect.GeneratedMethodAccessor88.invoke(Unknown Source)
     [exec] 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     [exec] 	at java.lang.reflect.Method.invoke(Method.java:498)
     [exec] 	at org.junit.internal.runners.TestMethod.invoke(TestMethod.java:59)
     [exec] 	at org.junit.internal.runners.MethodRoadie.runTestMethod(MethodRoadie.java:98)
     [exec] 	at org.junit.internal.runners.MethodRoadie$2.run(MethodRoadie.java:79)
     [exec] 	at org.junit.internal.runners.MethodRoadie.runBeforesThenTestThenAfters(MethodRoadie.java:87)
     [exec] 	at org.junit.internal.runners.MethodRoadie.runTest(MethodRoadie.java:77)
     [exec] 	at org.junit.internal.runners.MethodRoadie.run(MethodRoadie.java:42)
     [exec] 	at org.junit.internal.runners.JUnit4ClassRunner.invokeTestMethod(JUnit4ClassRunner.java:88)
     [exec] 	at org.junit.internal.runners.JUnit4ClassRunner.runMethods(JUnit4ClassRunner.java:51)
     [exec] 	at org.junit.runners.Parameterized$TestClassRunnerForParameters.run(Parameterized.java:98)
     [exec] 	at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
     [exec] 	at org.junit.runners.Parameterized.access$000(Parameterized.java:55)
     [exec] 	at org.junit.runners.Parameterized$1.run(Parameterized.java:131)
     [exec] 	at org.junit.internal.runners.ClassRoadie.runUnprotected(ClassRoadie.java:27)
     [exec] 	at org.junit.internal.runners.ClassRoadie.runProtected(ClassRoadie.java:37)
     [exec] 	at org.junit.runners.Parameterized.run(Parameterized.java:129)
     [exec] 	at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
     [exec] 	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:28)
     [exec] 	at org.junit.runner.JUnitCore.run(JUnitCore.java:130)
     [exec] 	at org.junit.runner.JUnitCore.run(JUnitCore.java:109)
     [exec] 	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
     [exec] 	at org.junit.runner.JUnitCore.runClasses(JUnitCore.java:60)
     [exec] 	at com.huawei.querymanagement.SelectQuerySuite.main(SelectQuerySuite.java:18)
     [exec] 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     [exec] 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
     [exec] 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     [exec] 	at java.lang.reflect.Method.invoke(Method.java:498)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:775)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
     [exec] 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
     [exec] 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)