You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@phoenix.apache.org by "Sandeep Guggilam (Jira)" <ji...@apache.org> on 2020/04/22 18:58:00 UTC

[jira] [Updated] (PHOENIX-5863) Upsert into view against a table with index throws exception when 4.14.3 client connects to 4.16 server

     [ https://issues.apache.org/jira/browse/PHOENIX-5863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Sandeep Guggilam updated PHOENIX-5863:
--------------------------------------
    Description: 
Steps to reproduce:
 # Bring up a 4.16 server ( current 4.x branch)
 # Connect a 4.14.3 client to the server and execute below queries
 # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 VARCHAR(10));
 # CREATE INDEX T1_IDX ON T1(V1);
 # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
 # UPSERT INTO T1_V1(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below exception***

*Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one index write failed after retries,serverTimestamp=1587513404084,*

*at org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*

*at org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*

*at org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*

*at org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*

*at org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*

*at org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*

*at org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*

*at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*

*at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1742)*

*at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1827)*

*at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1783)*

*at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1020)*

*at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3420)*

*at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3158)*

*at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3100)*

*at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:944)*

*at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:872)*

*at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2472)*

*at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36812)*

*at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2399)*

*at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)*

*at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:311)*

*at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:291)*

*Caused by: org.apache.phoenix.hbase.index.exception.MultiIndexWriteFailureException:  disableIndexOnFailure=false, Failed to write to multiple index tables: [V2#I2|#I2]*

*at org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:245)*

*... 21 more*

  was:
Steps to reproduce:
 # Bring up a 4.16 server ( current 4.x branch)
 # Connect a 4.14.3 client to the server and execute below queries
 # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 VARCHAR(10));
 # CREATE INDEX T1_IDX ON T1(V1);
 # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
 # UPSERT INTO T2_V2(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below exception***

*Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one index write failed after retries,serverTimestamp=1587513404084,*

 *at org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*

 *at org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*

 *at org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*

 *at org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*

 *at org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*

 *at org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*

 *at org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*

 *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*

 *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1742)*

 *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1827)*

 *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1783)*

 *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1020)*

 *at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3420)*

 *at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3158)*

 *at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3100)*

 *at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:944)*

 *at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:872)*

 *at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2472)*

 *at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36812)*

 *at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2399)*

 *at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)*

 *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:311)*

 *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:291)*

*Caused by: org.apache.phoenix.hbase.index.exception.MultiIndexWriteFailureException:  disableIndexOnFailure=false, Failed to write to multiple index tables: [V2#I2]*

 *at org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:245)*

 *... 21 more*


> Upsert into view against a table with index throws exception when 4.14.3 client connects to 4.16 server
> -------------------------------------------------------------------------------------------------------
>
>                 Key: PHOENIX-5863
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-5863
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.16.0
>            Reporter: Sandeep Guggilam
>            Priority: Blocker
>
> Steps to reproduce:
>  # Bring up a 4.16 server ( current 4.x branch)
>  # Connect a 4.14.3 client to the server and execute below queries
>  # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 VARCHAR(10));
>  # CREATE INDEX T1_IDX ON T1(V1);
>  # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
>  # UPSERT INTO T1_V1(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below exception***
> *Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one index write failed after retries,serverTimestamp=1587513404084,*
> *at org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*
> *at org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*
> *at org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*
> *at org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*
> *at org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*
> *at org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*
> *at org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*
> *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*
> *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1742)*
> *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1827)*
> *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1783)*
> *at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1020)*
> *at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3420)*
> *at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3158)*
> *at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3100)*
> *at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:944)*
> *at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:872)*
> *at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2472)*
> *at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36812)*
> *at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2399)*
> *at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)*
> *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:311)*
> *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:291)*
> *Caused by: org.apache.phoenix.hbase.index.exception.MultiIndexWriteFailureException:  disableIndexOnFailure=false, Failed to write to multiple index tables: [V2#I2|#I2]*
> *at org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:245)*
> *... 21 more*



--
This message was sent by Atlassian Jira
(v8.3.4#803005)