You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@phoenix.apache.org by "Thomas D'Silva (JIRA)" <ji...@apache.org> on 2019/01/08 23:04:00 UTC

[jira] [Updated] (PHOENIX-3204) Scanner lease timeout exception during UPSERT INTO t1 SELECT * FROM t2

     [ https://issues.apache.org/jira/browse/PHOENIX-3204?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Thomas D'Silva updated PHOENIX-3204:
------------------------------------
    Fix Version/s:     (was: 4.15.0)

> Scanner lease timeout exception during UPSERT INTO t1 SELECT * FROM t2
> ----------------------------------------------------------------------
>
>                 Key: PHOENIX-3204
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-3204
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.8.0
>            Reporter: Lars Hofhansl
>            Priority: Major
>         Attachments: PHOENIX-3204.patch
>
>
> We ran some larg'ish Phoenix tests on a larg'ish cluster.
> We had loaded 1bn rows into table t1 and tried to load the same data into a different table t2. The UPSERT failed about 1/2 way through with various exceptions like the following.
> Interesting are the exception in close() chain, as Phoenix could probably ignore those.
> Just filing here for reference.
> {code}
> 2016-08-23 00:56:41,851 INFO  [phoenix-1-thread-653] client.AsyncProcess - #4816618, waiting for some tasks to finish. Expected max=0, tasksSent=14, tasksDone=13, currentTasksDone=13, retries=13 hasError=false, tableName=null
> 2016-08-23 00:56:41,851 INFO  [phoenix-1-thread-653] client.AsyncProcess - Left over 1 task(s) are processed on server(s): [host17-8,60020,1471339710192]
> 2016-08-23 00:56:41,851 INFO  [phoenix-1-thread-653] client.AsyncProcess - Regions against which left over task(s) are processed: [TEST3,97432328$#}
> ,1471913672240.9f2fd6435921da1a203daf52f0116a34.]
> 2016-08-23 00:56:41,859 INFO  [phoenix-1-thread-293] client.AsyncProcess - #4816622, waiting for some tasks to finish. Expected max=0, tasksSent=14, tasksDone=13, currentTasksDone=13, retries=13 hasError=false, tableName=null
> 2016-08-23 00:56:41,859 INFO  [phoenix-1-thread-293] client.AsyncProcess - Left over 1 task(s) are processed on server(s): [host17-8,60020,1471339710192]
> 2016-08-23 00:56:41,859 INFO  [phoenix-1-thread-293] client.AsyncProcess - Regions against which left over task(s) are processed: [TEST3,97432328$#}
> ,1471913672240.9f2fd6435921da1a203daf52f0116a34.]
> 2016-08-23 00:56:41,860 INFO  [80-shared--pool2-t90] client.AsyncProcess - #4816618, table=TEST3, attempt=14/35 SUCCEEDED on host17-8,60020,1471339710192, tracking started Tue Aug 23 00:55:12 GMT 2016
> 2016-08-23 00:56:41,861 DEBUG [phoenix-1-thread-653] token.AuthenticationTokenSelector - No matching token found
> 2016-08-23 00:56:41,861 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - Creating SASL GSSAPI client. Server's Kerberos principal name is hbase/host28-20@
> 2016-08-23 00:56:41,862 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - Have sent token of size 730 from initSASLContext.
> 2016-08-23 00:56:41,863 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - Will read input token of size 108 for processing by initSASLContext
> 2016-08-23 00:56:41,863 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - Will send token of size 0 from initSASLContext.
> 2016-08-23 00:56:41,863 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - Will read input token of size 32 for processing by initSASLContext
> 2016-08-23 00:56:41,863 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - Will send token of size 32 from initSASLContext.
> 2016-08-23 00:56:41,863 DEBUG [phoenix-1-thread-653] security.HBaseSaslRpcClient - SASL client context established. Negotiated QoP: auth
> 2016-08-23 00:56:41,865 INFO  [phoenix-1-thread-653] client.ClientScanner - For hints related to the following exception, please try taking a look at: https://hbase.apache.org/book.html#trouble.client.scantimeout
> 2016-08-23 00:56:41,865 WARN  [phoenix-1-thread-653] client.ScannerCallable - Ignore, probably already closed
> org.apache.hadoop.hbase.UnknownScannerException: org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '49'. This can happen due to any of the following reasons: a) Scanner id given is wrong, b) Scanner lease expired because of long wait between consecutive client checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
> If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period' configuration.
> 	at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3228)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32492)
> 	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2208)
> 	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at sun.reflect.GeneratedConstructorAccessor19.newInstance(Unknown Source)
> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> 	at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> 	at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
> 	at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
> 	at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:305)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:309)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:164)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:58)
> 	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:115)
> 	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:91)
> 	at org.apache.hadoop.hbase.client.ClientScanner.close(ClientScanner.java:493)
> 	at org.apache.phoenix.iterate.ScanningResultIterator.close(ScanningResultIterator.java:49)
> 	at org.apache.phoenix.iterate.TableResultIterator.close(TableResultIterator.java:113)
> 	at org.apache.phoenix.jdbc.PhoenixResultSet.close(PhoenixResultSet.java:162)
> 	at org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:224)
> 	at org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:110)
> 	at org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:254)
> 	at org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
> 	at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:114)
> 	at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:106)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.UnknownScannerException): org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '49'. This can happen due to any of the following reasons: a) Scanner id given is wrong, b) Scanner lease expired because of long wait between consecutive client checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
> If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period' configuration.
> 	at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3228)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32492)
> 	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2208)
> 	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1512)
> 	at org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1714)
> 	at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1784)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:32938)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:307)
> 	... 19 more
> 2016-08-23 00:56:41,869 INFO  [80-shared--pool2-t15] client.AsyncProcess - #4816622, table=TEST3, attempt=14/35 SUCCEEDED on host17-8,60020,1471339710192, tracking started Tue Aug 23 00:55:12 GMT 2016
> 2016-08-23 00:56:41,869 INFO  [phoenix-1-thread-293] client.ClientScanner - For hints related to the following exception, please try taking a look at: https://hbase.apache.org/book.html#trouble.client.scantimeout
> 2016-08-23 00:56:41,870 WARN  [phoenix-1-thread-293] client.ScannerCallable - Ignore, probably already closed
> org.apache.hadoop.hbase.UnknownScannerException: org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '107'. This can happen due to any of the following reasons: a) Scanner id given is wrong, b) Scanner lease expired because of long wait between consecutive client checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
> If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period' configuration.
> 	at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3228)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32492)
> 	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2208)
> 	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at sun.reflect.GeneratedConstructorAccessor19.newInstance(Unknown Source)
> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> 	at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> 	at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
> 	at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
> 	at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:305)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:309)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:164)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:58)
> 	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:115)
> 	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:91)
> 	at org.apache.hadoop.hbase.client.ClientScanner.close(ClientScanner.java:493)
> 	at org.apache.phoenix.iterate.ScanningResultIterator.close(ScanningResultIterator.java:49)
> 	at org.apache.phoenix.iterate.TableResultIterator.close(TableResultIterator.java:113)
> 	at org.apache.phoenix.jdbc.PhoenixResultSet.close(PhoenixResultSet.java:162)
> 	at org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:224)
> 	at org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:110)
> 	at org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:254)
> 	at org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
> 	at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:114)
> 	at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:106)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.UnknownScannerException): org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '107'. This can happen due to any of the following reasons: a) Scanner id given is wrong, b) Scanner lease expired because of long wait between consecutive client checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
> If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period' configuration.
> 	at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3228)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32492)
> 	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2208)
> 	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1512)
> 	at org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1714)
> 	at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1784)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:32938)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:307)
> 	... 19 more
> 2016-08-23 00:56:41,890 INFO  [phoenix-1-thread-174] client.AsyncProcess - #4816630, waiting for some tasks to finish. Expected max=0, tasksSent=14, tasksDone=13, currentTasksDone=13, retries=13 hasError=false, tableName=null
> 2016-08-23 00:56:41,890 INFO  [phoenix-1-thread-174] client.AsyncProcess - Left over 1 task(s) are processed on server(s): [host17-8,60020,1471339710192]
> 2016-08-23 00:56:41,890 INFO  [phoenix-1-thread-174] client.AsyncProcess - Regions against which left over task(s) are processed: [TEST3,97432328$#}
> ,1471913672240.9f2fd6435921da1a203daf52f0116a34.]
> 2016-08-23 00:56:41,900 INFO  [0-shared--pool2-t149] client.AsyncProcess - #4816630, table=TEST3, attempt=14/35 SUCCEEDED on host17-8,60020,1471339710192, tracking started Tue Aug 23 00:55:12 GMT 2016
> 2016-08-23 00:56:41,901 DEBUG [phoenix-1-thread-174] token.AuthenticationTokenSelector - No matching token found
> 2016-08-23 00:56:41,901 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - Creating SASL GSSAPI client. Server's Kerberos principal name is hbase/host12-5@
> 2016-08-23 00:56:41,902 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - Have sent token of size 729 from initSASLContext.
> 2016-08-23 00:56:41,903 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - Will read input token of size 108 for processing by initSASLContext
> 2016-08-23 00:56:41,903 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - Will send token of size 0 from initSASLContext.
> 2016-08-23 00:56:41,903 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - Will read input token of size 32 for processing by initSASLContext
> 2016-08-23 00:56:41,903 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - Will send token of size 32 from initSASLContext.
> 2016-08-23 00:56:41,903 DEBUG [phoenix-1-thread-174] security.HBaseSaslRpcClient - SASL client context established. Negotiated QoP: auth
> 2016-08-23 00:56:41,904 INFO  [phoenix-1-thread-174] client.ClientScanner - For hints related to the following exception, please try taking a look at: https://hbase.apache.org/book.html#trouble.client.scantimeout
> 2016-08-23 00:56:41,905 WARN  [phoenix-1-thread-174] client.ScannerCallable - Ignore, probably already closed
> org.apache.hadoop.hbase.UnknownScannerException: org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '48'. This can happen due to any of the following reasons: a) Scanner id given is wrong, b) Scanner lease expired because of long wait between consecutive client checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
> If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period' configuration.
> 	at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3228)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32492)
> 	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2208)
> 	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at sun.reflect.GeneratedConstructorAccessor19.newInstance(Unknown Source)
> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> 	at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> 	at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
> 	at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
> 	at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:305)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:309)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:164)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:58)
> 	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:115)
> 	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:91)
> 	at org.apache.hadoop.hbase.client.ClientScanner.close(ClientScanner.java:493)
> 	at org.apache.phoenix.iterate.ScanningResultIterator.close(ScanningResultIterator.java:49)
> 	at org.apache.phoenix.iterate.TableResultIterator.close(TableResultIterator.java:113)
> 	at org.apache.phoenix.jdbc.PhoenixResultSet.close(PhoenixResultSet.java:162)
> 	at org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:224)
> 	at org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:110)
> 	at org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:254)
> 	at org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
> 	at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:114)
> 	at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:106)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.UnknownScannerException): org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '48'. This can happen due to any of the following reasons: a) Scanner id given is wrong, b) Scanner lease expired because of long wait between consecutive client checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
> If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period' configuration.
> 	at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3228)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32492)
> 	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2208)
> 	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
> 	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1512)
> 	at org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1714)
> 	at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1784)
> 	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:32938)
> 	at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:307)
> 	... 19 more
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)