You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by ps...@apache.org on 2019/05/10 08:37:20 UTC
[hbase] branch branch-2 updated: HBASE-21536 Fix completebulkload
usage instructions
This is an automated email from the ASF dual-hosted git repository.
psomogyi pushed a commit to branch branch-2
in repository https://gitbox.apache.org/repos/asf/hbase.git
The following commit(s) were added to refs/heads/branch-2 by this push:
new bbfb8ef HBASE-21536 Fix completebulkload usage instructions
bbfb8ef is described below
commit bbfb8ef80c7447bb0aa9616f6412b40c4e70837d
Author: Artem Ervits <ge...@gmail.com>
AuthorDate: Thu May 9 16:50:02 2019 -0400
HBASE-21536 Fix completebulkload usage instructions
Signed-off-by: Peter Somogyi <ps...@apache.org>
---
bin/hbase | 53 ++++++++++++----------
.../hadoop/hbase/tool/LoadIncrementalHFiles.java | 17 ++++---
src/main/asciidoc/_chapters/ops_mgt.adoc | 2 +-
3 files changed, 39 insertions(+), 33 deletions(-)
diff --git a/bin/hbase b/bin/hbase
index 8db26f1..1f16a32 100755
--- a/bin/hbase
+++ b/bin/hbase
@@ -91,33 +91,34 @@ if [ $# = 0 ]; then
echo ""
echo "Commands:"
echo "Some commands take arguments. Pass no args or -h for usage."
- echo " shell Run the HBase shell"
- echo " hbck Run the HBase 'fsck' tool. Defaults read-only hbck1."
- echo " Pass '-j /path/to/HBCK2.jar' to run hbase-2.x HBCK2."
- echo " snapshot Tool for managing snapshots"
+ echo " shell Run the HBase shell"
+ echo " hbck Run the HBase 'fsck' tool. Defaults read-only hbck1."
+ echo " Pass '-j /path/to/HBCK2.jar' to run hbase-2.x HBCK2."
+ echo " snapshot Tool for managing snapshots"
if [ "${in_omnibus_tarball}" = "true" ]; then
- echo " wal Write-ahead-log analyzer"
- echo " hfile Store file analyzer"
- echo " zkcli Run the ZooKeeper shell"
- echo " master Run an HBase HMaster node"
- echo " regionserver Run an HBase HRegionServer node"
- echo " zookeeper Run a ZooKeeper server"
- echo " rest Run an HBase REST server"
- echo " thrift Run the HBase Thrift server"
- echo " thrift2 Run the HBase Thrift2 server"
- echo " clean Run the HBase clean up script"
+ echo " wal Write-ahead-log analyzer"
+ echo " hfile Store file analyzer"
+ echo " zkcli Run the ZooKeeper shell"
+ echo " master Run an HBase HMaster node"
+ echo " regionserver Run an HBase HRegionServer node"
+ echo " zookeeper Run a ZooKeeper server"
+ echo " rest Run an HBase REST server"
+ echo " thrift Run the HBase Thrift server"
+ echo " thrift2 Run the HBase Thrift2 server"
+ echo " clean Run the HBase clean up script"
fi
- echo " classpath Dump hbase CLASSPATH"
- echo " mapredcp Dump CLASSPATH entries required by mapreduce"
- echo " pe Run PerformanceEvaluation"
- echo " ltt Run LoadTestTool"
- echo " canary Run the Canary tool"
- echo " version Print the version"
- echo " regionsplitter Run RegionSplitter tool"
- echo " rowcounter Run RowCounter tool"
- echo " cellcounter Run CellCounter tool"
- echo " pre-upgrade Run Pre-Upgrade validator tool"
- echo " CLASSNAME Run the class named CLASSNAME"
+ echo " classpath Dump hbase CLASSPATH"
+ echo " mapredcp Dump CLASSPATH entries required by mapreduce"
+ echo " pe Run PerformanceEvaluation"
+ echo " ltt Run LoadTestTool"
+ echo " canary Run the Canary tool"
+ echo " version Print the version"
+ echo " completebulkload Run BulkLoadHFiles tool"
+ echo " regionsplitter Run RegionSplitter tool"
+ echo " rowcounter Run RowCounter tool"
+ echo " cellcounter Run CellCounter tool"
+ echo " pre-upgrade Run Pre-Upgrade validator tool"
+ echo " CLASSNAME Run the class named CLASSNAME"
exit 1
fi
@@ -617,6 +618,8 @@ elif [ "$COMMAND" = "cellcounter" ] ; then
CLASS='org.apache.hadoop.hbase.mapreduce.CellCounter'
elif [ "$COMMAND" = "pre-upgrade" ] ; then
CLASS='org.apache.hadoop.hbase.tool.PreUpgradeValidator'
+elif [ "$COMMAND" = "completebulkload" ] ; then
+ CLASS='org.apache.hadoop.hbase.tool.BulkLoadHFilesTool'
else
CLASS=$COMMAND
fi
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.java
index dcb9ee5..fb87d79 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.java
@@ -183,13 +183,16 @@ public class LoadIncrementalHFiles extends Configured implements Tool {
}
private void usage() {
- System.err.println("usage: " + NAME + " /path/to/hfileoutputformat-output tablename -loadTable"
- + "\n -D" + CREATE_TABLE_CONF_KEY + "=no - can be used to avoid creation of table by "
- + "this tool\n Note: if you set this to 'no', then the target table must already exist "
- + "in HBase\n -loadTable implies your baseDirectory to store file has a depth of 3 ,you"
- + " must have an existing table\n-D" + IGNORE_UNMATCHED_CF_CONF_KEY + "=yes - can be used "
- + "to ignore unmatched column families\n" +
- "\n");
+ System.err.println("usage: " + "bin/hbase completebulkload <-Dargs> "
+ + "</path/to/hfileoutputformat-output> <tablename>\n"
+ + "\t-D" + CREATE_TABLE_CONF_KEY + "=no can be used to avoid creation "
+ + "of a table by this tool.\n"
+ + "\t Note: if you set this to 'no', then target table must already exist.\n"
+ + "\t-D" + IGNORE_UNMATCHED_CF_CONF_KEY + "=yes can be used to ignore "
+ + "unmatched column families.\n"
+ + "\t-loadTable switch implies your baseDirectory to store file has a "
+ + "depth of 3, table must exist\n"
+ + "\t and -loadTable switch is the last option on the command line.\n\n");
}
/**
diff --git a/src/main/asciidoc/_chapters/ops_mgt.adoc b/src/main/asciidoc/_chapters/ops_mgt.adoc
index e8414d3..54dd5ac 100644
--- a/src/main/asciidoc/_chapters/ops_mgt.adoc
+++ b/src/main/asciidoc/_chapters/ops_mgt.adoc
@@ -698,7 +698,7 @@ $ bin/hbase org.apache.hadoop.hbase.tool.LoadIncrementalHFiles <hdfs://storefile
.Driver
----
-HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` ${HADOOP_HOME}/bin/hadoop jar ${HBASE_HOME}/hbase-server-VERSION.jar completebulkload <hdfs://storefileoutput> <tablename>
+HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` ${HADOOP_HOME}/bin/hadoop jar ${HBASE_HOME}/hbase-mapreduce-VERSION.jar completebulkload <hdfs://storefileoutput> <tablename>
----
[[completebulkload.warning]]