You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Kumiko Yada <Ku...@ds-iq.com> on 2015/12/22 20:00:44 UTC
Put performance test
Hello,
I wrote the python script w/ happybase library to do the performance put test; however, this library is crashing when more than 900000 rows are put. I'd like to do 1/10/100 million rows put tests. Is there any tool that I can use for this?
Thanks
Kumiko
RE: Put performance test
Posted by Kumiko Yada <Ku...@ds-iq.com>.
I will try this. Thanks.
-----Original Message-----
From: Ted Yu [mailto:yuzhihong@gmail.com]
Sent: Tuesday, December 22, 2015 1:18 PM
To: user@hbase.apache.org
Subject: Re: Put performance test
Kumiko:
You can define your own YCSB workload by specifying the readproportion and scanproportion you want.
FYI
On Tue, Dec 22, 2015 at 11:39 AM, iain wright <ia...@gmail.com> wrote:
> You could use YCSB and a custom workload (i don't see a predefined
> workload for 100% puts without reads)
>
> https://github.com/brianfrankcooper/YCSB/wiki/Core-Workloads
>
> HBase also has a utility for running some evaluations via MR or a
> thread based client:
>
> $ ./hbase org.apache.hadoop.hbase.PerformanceEvaluation
> Usage: java org.apache.hadoop.hbase.PerformanceEvaluation \
> <OPTIONS> [-D<property=value>]* <command> <nclients>
>
> Options:
> nomapred Run multiple clients using threads (rather than use
> mapreduce)
> rows Rows each client runs. Default: One million
> size Total size in GiB. Mutually exclusive with --rows.
> Default: 1.0.
> sampleRate Execute test on a sample of total rows. Only supported by
> randomRead. Default: 1.0
> traceRate Enable HTrace spans. Initiate tracing every N rows.
> Default: 0
> table Alternate table name. Default: 'TestTable'
> multiGet If >0, when doing RandomRead, perform multiple gets
> instead of single gets. Default: 0
> compress Compression type to use (GZ, LZO, ...). Default: 'NONE'
> flushCommits Used to determine if the test should flush the table.
> Default: false
> writeToWAL Set writeToWAL on puts. Default: True
> autoFlush Set autoFlush on htable. Default: False
> oneCon all the threads share the same connection. Default: False
> presplit Create presplit table. Recommended for accurate perf
> analysis (see guide). Default: disabled
> inmemory Tries to keep the HFiles of the CF inmemory as far as
> possible. Not guaranteed that reads are always served from memory.
> Default: false
> usetags Writes tags along with KVs. Use with HFile V3. Default:
> false
> numoftags Specify the no of tags that would be needed. This works
> only if usetags is true.
> filterAll Helps to filter out all the rows on the server side there
> by not returning any thing back to the client. Helps to check the
> server side performance. Uses FilterAllFilter internally.
> latency Set to report operation latencies. Default: False
> bloomFilter Bloom filter type, one of [NONE, ROW, ROWCOL]
> valueSize Pass value size to use: Default: 1024
> valueRandom Set if we should vary value size between 0 and
> 'valueSize'; set on read for stats on size: Default: Not set.
> valueZipf Set if we should vary value size between 0 and 'valueSize'
> in zipf form: Default: Not set.
> period Report every 'period' rows: Default: opts.perClientRunRows
> / 10
> multiGet Batch gets together into groups of N. Only supported by
> randomRead. Default: disabled
> addColumns Adds columns to scans/gets explicitly. Default: true
> replicas Enable region replica testing. Defaults: 1.
> splitPolicy Specify a custom RegionSplitPolicy for the table.
> randomSleep Do a random sleep before each get between 0 and entered
> value. Defaults: 0
> columns Columns to write per row. Default: 1
> caching Scan caching to use. Default: 30
>
> Note: -D properties will be applied to the conf used.
> For example:
> -Dmapreduce.output.fileoutputformat.compress=true
> -Dmapreduce.task.timeout=60000
>
> Command:
> filterScan Run scan test using a filter to find a specific row based
> on it's value (make sure to use --rows=20)
> randomRead Run random read test
> randomSeekScan Run random seek and scan 100 test
> randomWrite Run random write test
> scan Run scan test (read every row)
> scanRange10 Run random seek scan with both start and stop row (max 10
> rows)
> scanRange100 Run random seek scan with both start and stop row (max 100
> rows)
> scanRange1000 Run random seek scan with both start and stop row (max
> 1000 rows)
> scanRange10000 Run random seek scan with both start and stop row
> (max
> 10000 rows)
> sequentialRead Run sequential read test sequentialWrite Run
> sequential write test
>
> Args:
> nclients Integer. Required. Total number of clients (and
> HRegionServers)
> running: 1 <= value <= 500
> Examples:
> To run a single evaluation client:
> $ bin/hbase org.apache.hadoop.hbase.PerformanceEvaluation
> sequentialWrite
> 1
>
>
>
> --
> Iain Wright
>
> This email message is confidential, intended only for the recipient(s)
> named above and may contain information that is privileged, exempt
> from disclosure under applicable law. If you are not the intended
> recipient, do not disclose or disseminate the message to anyone except
> the intended recipient. If you have received this message in error, or
> are not the named recipient(s), please immediately notify the sender
> by return email, and delete all copies of this message.
>
> On Tue, Dec 22, 2015 at 11:12 AM, Kumiko Yada <Ku...@ds-iq.com>
> wrote:
>
> > For to add that I don't want to the bulkinsert for this test.
> >
> > Thanks
> > Kumiko
> >
> > -----Original Message-----
> > From: Kumiko Yada [mailto:Kumiko.Yada@ds-iq.com]
> > Sent: Tuesday, December 22, 2015 11:01 AM
> > To: user@hbase.apache.org
> > Subject: Put performance test
> >
> > Hello,
> >
> > I wrote the python script w/ happybase library to do the performance
> > put test; however, this library is crashing when more than 900000
> > rows are put. I'd like to do 1/10/100 million rows put tests. Is
> > there any tool that I can use for this?
> >
> > Thanks
> > Kumiko
> >
>
Re: Put performance test
Posted by Ted Yu <yu...@gmail.com>.
Kumiko:
You can define your own YCSB workload by specifying the readproportion
and scanproportion you want.
FYI
On Tue, Dec 22, 2015 at 11:39 AM, iain wright <ia...@gmail.com> wrote:
> You could use YCSB and a custom workload (i don't see a predefined workload
> for 100% puts without reads)
>
> https://github.com/brianfrankcooper/YCSB/wiki/Core-Workloads
>
> HBase also has a utility for running some evaluations via MR or a thread
> based client:
>
> $ ./hbase org.apache.hadoop.hbase.PerformanceEvaluation
> Usage: java org.apache.hadoop.hbase.PerformanceEvaluation \
> <OPTIONS> [-D<property=value>]* <command> <nclients>
>
> Options:
> nomapred Run multiple clients using threads (rather than use
> mapreduce)
> rows Rows each client runs. Default: One million
> size Total size in GiB. Mutually exclusive with --rows.
> Default: 1.0.
> sampleRate Execute test on a sample of total rows. Only supported by
> randomRead. Default: 1.0
> traceRate Enable HTrace spans. Initiate tracing every N rows.
> Default: 0
> table Alternate table name. Default: 'TestTable'
> multiGet If >0, when doing RandomRead, perform multiple gets
> instead of single gets. Default: 0
> compress Compression type to use (GZ, LZO, ...). Default: 'NONE'
> flushCommits Used to determine if the test should flush the table.
> Default: false
> writeToWAL Set writeToWAL on puts. Default: True
> autoFlush Set autoFlush on htable. Default: False
> oneCon all the threads share the same connection. Default: False
> presplit Create presplit table. Recommended for accurate perf
> analysis (see guide). Default: disabled
> inmemory Tries to keep the HFiles of the CF inmemory as far as
> possible. Not guaranteed that reads are always served from memory.
> Default: false
> usetags Writes tags along with KVs. Use with HFile V3. Default:
> false
> numoftags Specify the no of tags that would be needed. This works
> only if usetags is true.
> filterAll Helps to filter out all the rows on the server side there
> by not returning any thing back to the client. Helps to check the server
> side performance. Uses FilterAllFilter internally.
> latency Set to report operation latencies. Default: False
> bloomFilter Bloom filter type, one of [NONE, ROW, ROWCOL]
> valueSize Pass value size to use: Default: 1024
> valueRandom Set if we should vary value size between 0 and
> 'valueSize'; set on read for stats on size: Default: Not set.
> valueZipf Set if we should vary value size between 0 and 'valueSize'
> in zipf form: Default: Not set.
> period Report every 'period' rows: Default: opts.perClientRunRows
> / 10
> multiGet Batch gets together into groups of N. Only supported by
> randomRead. Default: disabled
> addColumns Adds columns to scans/gets explicitly. Default: true
> replicas Enable region replica testing. Defaults: 1.
> splitPolicy Specify a custom RegionSplitPolicy for the table.
> randomSleep Do a random sleep before each get between 0 and entered
> value. Defaults: 0
> columns Columns to write per row. Default: 1
> caching Scan caching to use. Default: 30
>
> Note: -D properties will be applied to the conf used.
> For example:
> -Dmapreduce.output.fileoutputformat.compress=true
> -Dmapreduce.task.timeout=60000
>
> Command:
> filterScan Run scan test using a filter to find a specific row based
> on it's value (make sure to use --rows=20)
> randomRead Run random read test
> randomSeekScan Run random seek and scan 100 test
> randomWrite Run random write test
> scan Run scan test (read every row)
> scanRange10 Run random seek scan with both start and stop row (max 10
> rows)
> scanRange100 Run random seek scan with both start and stop row (max 100
> rows)
> scanRange1000 Run random seek scan with both start and stop row (max
> 1000 rows)
> scanRange10000 Run random seek scan with both start and stop row (max
> 10000 rows)
> sequentialRead Run sequential read test
> sequentialWrite Run sequential write test
>
> Args:
> nclients Integer. Required. Total number of clients (and
> HRegionServers)
> running: 1 <= value <= 500
> Examples:
> To run a single evaluation client:
> $ bin/hbase org.apache.hadoop.hbase.PerformanceEvaluation sequentialWrite
> 1
>
>
>
> --
> Iain Wright
>
> This email message is confidential, intended only for the recipient(s)
> named above and may contain information that is privileged, exempt from
> disclosure under applicable law. If you are not the intended recipient, do
> not disclose or disseminate the message to anyone except the intended
> recipient. If you have received this message in error, or are not the named
> recipient(s), please immediately notify the sender by return email, and
> delete all copies of this message.
>
> On Tue, Dec 22, 2015 at 11:12 AM, Kumiko Yada <Ku...@ds-iq.com>
> wrote:
>
> > For to add that I don't want to the bulkinsert for this test.
> >
> > Thanks
> > Kumiko
> >
> > -----Original Message-----
> > From: Kumiko Yada [mailto:Kumiko.Yada@ds-iq.com]
> > Sent: Tuesday, December 22, 2015 11:01 AM
> > To: user@hbase.apache.org
> > Subject: Put performance test
> >
> > Hello,
> >
> > I wrote the python script w/ happybase library to do the performance put
> > test; however, this library is crashing when more than 900000 rows are
> > put. I'd like to do 1/10/100 million rows put tests. Is there any tool
> > that I can use for this?
> >
> > Thanks
> > Kumiko
> >
>
Re: Put performance test
Posted by iain wright <ia...@gmail.com>.
You could use YCSB and a custom workload (i don't see a predefined workload
for 100% puts without reads)
https://github.com/brianfrankcooper/YCSB/wiki/Core-Workloads
HBase also has a utility for running some evaluations via MR or a thread
based client:
$ ./hbase org.apache.hadoop.hbase.PerformanceEvaluation
Usage: java org.apache.hadoop.hbase.PerformanceEvaluation \
<OPTIONS> [-D<property=value>]* <command> <nclients>
Options:
nomapred Run multiple clients using threads (rather than use
mapreduce)
rows Rows each client runs. Default: One million
size Total size in GiB. Mutually exclusive with --rows.
Default: 1.0.
sampleRate Execute test on a sample of total rows. Only supported by
randomRead. Default: 1.0
traceRate Enable HTrace spans. Initiate tracing every N rows.
Default: 0
table Alternate table name. Default: 'TestTable'
multiGet If >0, when doing RandomRead, perform multiple gets
instead of single gets. Default: 0
compress Compression type to use (GZ, LZO, ...). Default: 'NONE'
flushCommits Used to determine if the test should flush the table.
Default: false
writeToWAL Set writeToWAL on puts. Default: True
autoFlush Set autoFlush on htable. Default: False
oneCon all the threads share the same connection. Default: False
presplit Create presplit table. Recommended for accurate perf
analysis (see guide). Default: disabled
inmemory Tries to keep the HFiles of the CF inmemory as far as
possible. Not guaranteed that reads are always served from memory.
Default: false
usetags Writes tags along with KVs. Use with HFile V3. Default:
false
numoftags Specify the no of tags that would be needed. This works
only if usetags is true.
filterAll Helps to filter out all the rows on the server side there
by not returning any thing back to the client. Helps to check the server
side performance. Uses FilterAllFilter internally.
latency Set to report operation latencies. Default: False
bloomFilter Bloom filter type, one of [NONE, ROW, ROWCOL]
valueSize Pass value size to use: Default: 1024
valueRandom Set if we should vary value size between 0 and
'valueSize'; set on read for stats on size: Default: Not set.
valueZipf Set if we should vary value size between 0 and 'valueSize'
in zipf form: Default: Not set.
period Report every 'period' rows: Default: opts.perClientRunRows
/ 10
multiGet Batch gets together into groups of N. Only supported by
randomRead. Default: disabled
addColumns Adds columns to scans/gets explicitly. Default: true
replicas Enable region replica testing. Defaults: 1.
splitPolicy Specify a custom RegionSplitPolicy for the table.
randomSleep Do a random sleep before each get between 0 and entered
value. Defaults: 0
columns Columns to write per row. Default: 1
caching Scan caching to use. Default: 30
Note: -D properties will be applied to the conf used.
For example:
-Dmapreduce.output.fileoutputformat.compress=true
-Dmapreduce.task.timeout=60000
Command:
filterScan Run scan test using a filter to find a specific row based
on it's value (make sure to use --rows=20)
randomRead Run random read test
randomSeekScan Run random seek and scan 100 test
randomWrite Run random write test
scan Run scan test (read every row)
scanRange10 Run random seek scan with both start and stop row (max 10
rows)
scanRange100 Run random seek scan with both start and stop row (max 100
rows)
scanRange1000 Run random seek scan with both start and stop row (max
1000 rows)
scanRange10000 Run random seek scan with both start and stop row (max
10000 rows)
sequentialRead Run sequential read test
sequentialWrite Run sequential write test
Args:
nclients Integer. Required. Total number of clients (and
HRegionServers)
running: 1 <= value <= 500
Examples:
To run a single evaluation client:
$ bin/hbase org.apache.hadoop.hbase.PerformanceEvaluation sequentialWrite 1
--
Iain Wright
This email message is confidential, intended only for the recipient(s)
named above and may contain information that is privileged, exempt from
disclosure under applicable law. If you are not the intended recipient, do
not disclose or disseminate the message to anyone except the intended
recipient. If you have received this message in error, or are not the named
recipient(s), please immediately notify the sender by return email, and
delete all copies of this message.
On Tue, Dec 22, 2015 at 11:12 AM, Kumiko Yada <Ku...@ds-iq.com> wrote:
> For to add that I don't want to the bulkinsert for this test.
>
> Thanks
> Kumiko
>
> -----Original Message-----
> From: Kumiko Yada [mailto:Kumiko.Yada@ds-iq.com]
> Sent: Tuesday, December 22, 2015 11:01 AM
> To: user@hbase.apache.org
> Subject: Put performance test
>
> Hello,
>
> I wrote the python script w/ happybase library to do the performance put
> test; however, this library is crashing when more than 900000 rows are
> put. I'd like to do 1/10/100 million rows put tests. Is there any tool
> that I can use for this?
>
> Thanks
> Kumiko
>
RE: Put performance test
Posted by Kumiko Yada <Ku...@ds-iq.com>.
For to add that I don't want to the bulkinsert for this test.
Thanks
Kumiko
-----Original Message-----
From: Kumiko Yada [mailto:Kumiko.Yada@ds-iq.com]
Sent: Tuesday, December 22, 2015 11:01 AM
To: user@hbase.apache.org
Subject: Put performance test
Hello,
I wrote the python script w/ happybase library to do the performance put test; however, this library is crashing when more than 900000 rows are put. I'd like to do 1/10/100 million rows put tests. Is there any tool that I can use for this?
Thanks
Kumiko
RE: Put performance test
Posted by Frank Luo <jl...@merkleinc.com>.
If the table is not pre-split properly and timeout not increased, then region server will crash when compacting.
-----Original Message-----
From: Kumiko Yada [mailto:Kumiko.Yada@ds-iq.com]
Sent: Tuesday, December 22, 2015 1:01 PM
To: user@hbase.apache.org
Subject: Put performance test
Hello,
I wrote the python script w/ happybase library to do the performance put test; however, this library is crashing when more than 900000 rows are put. I'd like to do 1/10/100 million rows put tests. Is there any tool that I can use for this?
Thanks
Kumiko
This email and any attachments transmitted with it are intended for use by the intended recipient(s) only. If you have received this email in error, please notify the sender immediately and then delete it. If you are not the intended recipient, you must not keep, use, disclose, copy or distribute this email without the author’s prior permission. We take precautions to minimize the risk of transmitting software viruses, but we advise you to perform your own virus checks on any attachment to this message. We cannot accept liability for any loss or damage caused by software viruses. The information contained in this communication may be confidential and may be subject to the attorney-client privilege.
Re: Put performance test
Posted by Stack <st...@duboce.net>.
Looks like happybase has tuned down the retries. Is that possible? It
says: Failed
1 action: IOException: 1 time. Can you up the retry count? Client probably
needs to ride over a region split or some such.
St.Ack
On Wed, Dec 23, 2015 at 9:47 AM, Kumiko Yada <Ku...@ds-iq.com> wrote:
> It's crashing in the lib.
>
> happybase.hbase.ttypes.IOError:
> IOError(_message='org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException:
> Failed 1 action: IOException: 1 time, \n\tat
> org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.makeException(AsyncProcess.java:227)\n\tat
> org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.access$1700(AsyncProcess.java:207)\n\tat
> org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1658)\n\tat
> org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:208)\n\tat
> org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:183)\n\tat
> org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1496)\n\tat
> org.apache.hadoop.hbase.client.HTable.put(HTable.java:1119)\n\tat
> org.apache.hadoop.hbase.thrift.ThriftServerRunner$HBaseHandler.mutateRowsTs(ThriftServerRunner.java:1324)\n\tat
> org.apache.hadoop.hbase.thrift.ThriftServerRunner$HBaseHandler.mutateRows(ThriftServerRunner.java:1270)\n\tat
> sun.reflect.GeneratedMethodAccessor1.invoke(Unknown Source)\n\tat
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)\n\tat
> java.lang.reflect.Method.invoke(Method.java:606)\n\tat
> org.apache.hadoop.hbase.thrift.HbaseHandlerMetricsProxy.invoke(HbaseHandlerMetricsProxy.java:67)\n\tat
> com.sun.proxy.$Proxy13.mutateRows(Unknown Source)\n\tat
> org.apache.hadoop.hbase.thrift.generated.Hbase$Processor$mutateRows.getResult(Hbase.java:4386)\n\tat
> org.apache.hadoop.hbase.thrift.generated.Hbase$Processor$mutateRows.getResult(Hbase.java:4370)\n\tat
> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)\n\tat
> org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)\n\tat
> org.apache.hadoop.hbase.thrift.TBoundedThreadPoolServer$ClientConnnection.run(TBoundedThreadPoolServer.java:289)\n\tat
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)\n\tat
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)\n\tat
> java.lang.Thread.run(Thread.java:745)\n')
>
> -----Original Message-----
> From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of Stack
> Sent: Tuesday, December 22, 2015 1:07 PM
> To: Hbase-User <us...@hbase.apache.org>
> Subject: Re: Put performance test
>
> Why the crash? Is it in the library or hbase?
> Thanks,
> St.Ack
>
> On Tue, Dec 22, 2015 at 11:00 AM, Kumiko Yada <Ku...@ds-iq.com>
> wrote:
>
> > Hello,
> >
> > I wrote the python script w/ happybase library to do the performance
> > put test; however, this library is crashing when more than 900000 rows
> > are put. I'd like to do 1/10/100 million rows put tests. Is there
> > any tool that I can use for this?
> >
> > Thanks
> > Kumiko
> >
>
RE: Put performance test
Posted by Kumiko Yada <Ku...@ds-iq.com>.
It's crashing in the lib.
happybase.hbase.ttypes.IOError: IOError(_message='org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: IOException: 1 time, \n\tat org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.makeException(AsyncProcess.java:227)\n\tat org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.access$1700(AsyncProcess.java:207)\n\tat org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1658)\n\tat org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:208)\n\tat org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:183)\n\tat org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1496)\n\tat org.apache.hadoop.hbase.client.HTable.put(HTable.java:1119)\n\tat org.apache.hadoop.hbase.thrift.ThriftServerRunner$HBaseHandler.mutateRowsTs(ThriftServerRunner.java:1324)\n\tat org.apache.hadoop.hbase.thrift.ThriftServerRunner$HBaseHandler.mutateRows(ThriftServerRunner.java:1270)\n\tat sun.reflect.GeneratedMethodAccessor1.invoke(Unknown Source)\n\tat sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)\n\tat java.lang.reflect.Method.invoke(Method.java:606)\n\tat org.apache.hadoop.hbase.thrift.HbaseHandlerMetricsProxy.invoke(HbaseHandlerMetricsProxy.java:67)\n\tat com.sun.proxy.$Proxy13.mutateRows(Unknown Source)\n\tat org.apache.hadoop.hbase.thrift.generated.Hbase$Processor$mutateRows.getResult(Hbase.java:4386)\n\tat org.apache.hadoop.hbase.thrift.generated.Hbase$Processor$mutateRows.getResult(Hbase.java:4370)\n\tat org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)\n\tat org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)\n\tat org.apache.hadoop.hbase.thrift.TBoundedThreadPoolServer$ClientConnnection.run(TBoundedThreadPoolServer.java:289)\n\tat java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)\n\tat java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)\n\tat java.lang.Thread.run(Thread.java:745)\n')
-----Original Message-----
From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of Stack
Sent: Tuesday, December 22, 2015 1:07 PM
To: Hbase-User <us...@hbase.apache.org>
Subject: Re: Put performance test
Why the crash? Is it in the library or hbase?
Thanks,
St.Ack
On Tue, Dec 22, 2015 at 11:00 AM, Kumiko Yada <Ku...@ds-iq.com> wrote:
> Hello,
>
> I wrote the python script w/ happybase library to do the performance
> put test; however, this library is crashing when more than 900000 rows
> are put. I'd like to do 1/10/100 million rows put tests. Is there
> any tool that I can use for this?
>
> Thanks
> Kumiko
>
Re: Put performance test
Posted by Stack <st...@duboce.net>.
Why the crash? Is it in the library or hbase?
Thanks,
St.Ack
On Tue, Dec 22, 2015 at 11:00 AM, Kumiko Yada <Ku...@ds-iq.com> wrote:
> Hello,
>
> I wrote the python script w/ happybase library to do the performance put
> test; however, this library is crashing when more than 900000 rows are
> put. I'd like to do 1/10/100 million rows put tests. Is there any tool
> that I can use for this?
>
> Thanks
> Kumiko
>