You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by te...@apache.org on 2018/08/15 15:30:34 UTC
hbase git commit: HBASE-21040 Replace call to printStackTrace() with
proper logger call
Repository: hbase
Updated Branches:
refs/heads/master 1114a1a65 -> 49ae8549c
HBASE-21040 Replace call to printStackTrace() with proper logger call
Signed-off-by: tedyu <yu...@gmail.com>
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/49ae8549
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/49ae8549
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/49ae8549
Branch: refs/heads/master
Commit: 49ae8549cfcb30d56cd6342c14e6ad574e1c6383
Parents: 1114a1a
Author: subrat.mishra <su...@yahoo.com>
Authored: Wed Aug 15 10:38:51 2018 +0530
Committer: tedyu <yu...@gmail.com>
Committed: Wed Aug 15 08:29:54 2018 -0700
----------------------------------------------------------------------
.../hadoop/hbase/backup/RestoreDriver.java | 2 +-
.../client/ClientAsyncPrefetchScanner.java | 1 -
.../RpcRetryingCallerWithReadReplicas.java | 1 -
.../apache/hadoop/hbase/filter/ParseFilter.java | 23 ++++++++++----------
.../hadoop/hbase/mapreduce/SampleUploader.java | 6 ++++-
.../apache/hadoop/hbase/thrift/DemoClient.java | 19 +++++++++-------
.../hadoop/hbase/thrift/HttpDoAsClient.java | 7 ++++--
.../hadoop/hbase/mapred/TableMapReduceUtil.java | 9 +++++---
.../hadoop/hbase/mapreduce/CellCounter.java | 3 ++-
.../hadoop/hbase/mapreduce/CopyTable.java | 2 +-
.../hadoop/hbase/mapreduce/HashTable.java | 2 +-
.../apache/hadoop/hbase/mapreduce/Import.java | 9 +++++---
.../hadoop/hbase/mapreduce/SyncTable.java | 2 +-
.../hbase/mapreduce/TsvImporterMapper.java | 9 +++++---
.../hbase/mapreduce/TsvImporterTextMapper.java | 8 ++++---
.../hadoop/hbase/mapreduce/WALPlayer.java | 6 +++--
.../replication/VerifyReplication.java | 2 +-
.../store/wal/ProcedureWALPrettyPrinter.java | 6 ++++-
.../apache/hadoop/hbase/LocalHBaseCluster.java | 7 ++++--
.../hadoop/hbase/master/HMasterCommandLine.java | 2 +-
.../hadoop/hbase/wal/WALPrettyPrinter.java | 6 ++++-
.../hadoop/hbase/thrift/ThriftServer.java | 2 +-
.../hadoop/hbase/zookeeper/HQuorumPeer.java | 6 ++++-
23 files changed, 89 insertions(+), 51 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java
----------------------------------------------------------------------
diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java
index 4e90563..41ce5d4 100644
--- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java
+++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/RestoreDriver.java
@@ -180,7 +180,7 @@ public class RestoreDriver extends AbstractHBaseTool {
client.restore(BackupUtils.createRestoreRequest(backupRootDir, backupId, check,
sTableArray, tTableArray, overwrite));
} catch (Exception e) {
- e.printStackTrace();
+ LOG.error("Error while running restore backup", e);
return -5;
}
return 0;
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientAsyncPrefetchScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientAsyncPrefetchScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientAsyncPrefetchScanner.java
index 4529cf5..e5af871 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientAsyncPrefetchScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientAsyncPrefetchScanner.java
@@ -146,7 +146,6 @@ public class ClientAsyncPrefetchScanner extends ClientSimpleScanner {
// Rethrow the exception so the application can handle it.
while (!exceptionsQueue.isEmpty()) {
Exception first = exceptionsQueue.peek();
- first.printStackTrace();
if (first instanceof IOException) {
throw (IOException) first;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
index a0be0bf..4a31cff 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
@@ -279,7 +279,6 @@ public class RpcRetryingCallerWithReadReplicas {
throws RetriesExhaustedException, DoNotRetryIOException {
Throwable t = e.getCause();
assert t != null; // That's what ExecutionException is about: holding an exception
- t.printStackTrace();
if (t instanceof RetriesExhaustedException) {
throw (RetriesExhaustedException) t;
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
index 716322c..5428ed8 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
@@ -244,27 +244,28 @@ public class ParseFilter {
throws CharacterCodingException {
String filterName = Bytes.toString(getFilterName(filterStringAsByteArray));
- ArrayList<byte []> filterArguments = getFilterArguments(filterStringAsByteArray);
+ ArrayList<byte[]> filterArguments = getFilterArguments(filterStringAsByteArray);
if (!filterHashMap.containsKey(filterName)) {
throw new IllegalArgumentException("Filter Name " + filterName + " not supported");
}
+ filterName = filterHashMap.get(filterName);
+ final String methodName = "createFilterFromArguments";
try {
- filterName = filterHashMap.get(filterName);
Class<?> c = Class.forName(filterName);
- Class<?>[] argTypes = new Class [] {ArrayList.class};
- Method m = c.getDeclaredMethod("createFilterFromArguments", argTypes);
- return (Filter) m.invoke(null,filterArguments);
+ Class<?>[] argTypes = new Class[] { ArrayList.class };
+ Method m = c.getDeclaredMethod(methodName, argTypes);
+ return (Filter) m.invoke(null, filterArguments);
} catch (ClassNotFoundException e) {
- e.printStackTrace();
+ LOG.error("Could not find class {}", filterName, e);
} catch (NoSuchMethodException e) {
- e.printStackTrace();
+ LOG.error("Could not find method {} in {}", methodName, filterName, e);
} catch (IllegalAccessException e) {
- e.printStackTrace();
+ LOG.error("Unable to access specified class {}", filterName, e);
} catch (InvocationTargetException e) {
- e.printStackTrace();
+ LOG.error("Method {} threw an exception for {}", methodName, filterName, e);
}
- throw new IllegalArgumentException("Incorrect filter string " +
- new String(filterStringAsByteArray, StandardCharsets.UTF_8));
+ throw new IllegalArgumentException(
+ "Incorrect filter string " + new String(filterStringAsByteArray, StandardCharsets.UTF_8));
}
/**
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java
index 1248f87..3273f49 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/SampleUploader.java
@@ -35,6 +35,8 @@ import org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
import org.apache.yetus.audience.InterfaceAudience;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* Sample Uploader MapReduce
@@ -60,6 +62,7 @@ import org.apache.yetus.audience.InterfaceAudience;
*/
@InterfaceAudience.Private
public class SampleUploader extends Configured implements Tool {
+ private static final Logger LOG = LoggerFactory.getLogger(SampleUploader.class);
private static final String NAME = "SampleUploader";
@@ -100,7 +103,8 @@ public class SampleUploader extends Configured implements Tool {
try {
context.write(new ImmutableBytesWritable(row), put);
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted emitting put", e);
+ Thread.currentThread().interrupt();
}
// Set status every checkpoint lines
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
index 07486bf..4c3a99c 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
@@ -48,12 +48,15 @@ import org.apache.thrift.transport.TSaslClientTransport;
import org.apache.thrift.transport.TSocket;
import org.apache.thrift.transport.TTransport;
import org.apache.yetus.audience.InterfaceAudience;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* See the instructions under hbase-examples/README.txt
*/
@InterfaceAudience.Private
public class DemoClient {
+ private static final Logger LOG = LoggerFactory.getLogger(DemoClient.class);
static protected int port;
static protected String host;
@@ -110,15 +113,15 @@ public class DemoClient {
}
}
- // Helper to translate strings to UTF8 bytes
- private byte[] bytes(String s) {
- try {
- return s.getBytes("UTF-8");
- } catch (UnsupportedEncodingException e) {
- e.printStackTrace();
- return null;
- }
+ // Helper to translate strings to UTF8 bytes
+ private byte[] bytes(String s) {
+ try {
+ return s.getBytes("UTF-8");
+ } catch (UnsupportedEncodingException e) {
+ LOG.error("CharSetName {} not supported", s, e);
+ return null;
}
+ }
private void run() throws Exception {
TTransport transport = new TSocket(host, port);
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
index 422d405..fd214d1 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
@@ -53,12 +53,15 @@ import org.ietf.jgss.GSSException;
import org.ietf.jgss.GSSManager;
import org.ietf.jgss.GSSName;
import org.ietf.jgss.Oid;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* See the instructions under hbase-examples/README.txt
*/
@InterfaceAudience.Private
public class HttpDoAsClient {
+ private static final Logger LOG = LoggerFactory.getLogger(HttpDoAsClient.class);
static protected int port;
static protected String host;
@@ -113,7 +116,7 @@ public class HttpDoAsClient {
try {
return s.getBytes("UTF-8");
} catch (UnsupportedEncodingException e) {
- e.printStackTrace();
+ LOG.error("CharSetName {} not supported", s, e);
return null;
}
}
@@ -188,7 +191,7 @@ public class HttpDoAsClient {
try {
httpClient.setCustomHeader("Authorization", generateTicket());
} catch (GSSException e) {
- e.printStackTrace();
+ LOG.error("Kerberos authentication failed", e);
}
}
return client;
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java
index 0427f50..0907fd4 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java
@@ -40,6 +40,8 @@ import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.OutputFormat;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapred.TextOutputFormat;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import java.io.IOException;
import java.util.Collection;
@@ -51,6 +53,7 @@ import java.util.Map;
@InterfaceAudience.Public
@SuppressWarnings({ "rawtypes", "unchecked" })
public class TableMapReduceUtil {
+ private static final Logger LOG = LoggerFactory.getLogger(TableMapReduceUtil.class);
/**
* Use this before submitting a TableMap job. It will
@@ -110,14 +113,14 @@ public class TableMapReduceUtil {
try {
addDependencyJars(job);
} catch (IOException e) {
- e.printStackTrace();
+ LOG.error("IOException encountered while adding dependency jars", e);
}
}
try {
initCredentials(job);
} catch (IOException ioe) {
// just spit out the stack trace? really?
- ioe.printStackTrace();
+ LOG.error("IOException encountered while initializing credentials", ioe);
}
}
@@ -310,7 +313,7 @@ public class TableMapReduceUtil {
User user = userProvider.getCurrent();
TokenUtil.addTokenForJob(conn, job, user);
} catch (InterruptedException ie) {
- ie.printStackTrace();
+ LOG.error("Interrupted obtaining user authentication token", ie);
Thread.currentThread().interrupt();
} finally {
conn.close();
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java
index ff0f01c..f00a0f0 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java
@@ -176,7 +176,8 @@ public class CellCounter extends Configured implements Tool {
context.getCounter(Counters.CELLS).increment(cellCount);
}
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while writing cellCount", e);
+ Thread.currentThread().interrupt();
}
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java
index 2e9e62c..b20f07f 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java
@@ -339,7 +339,7 @@ public class CopyTable extends Configured implements Tool {
dstTableName = tableName;
}
} catch (Exception e) {
- e.printStackTrace();
+ LOG.error("Failed to parse commandLine arguments", e);
printUsage("Can't start because " + e.getMessage());
return false;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HashTable.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HashTable.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HashTable.java
index b67225e..2d742fc 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HashTable.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HashTable.java
@@ -713,7 +713,7 @@ public class HashTable extends Configured implements Tool {
}
} catch (Exception e) {
- e.printStackTrace();
+ LOG.error("Failed to parse commandLine arguments", e);
printUsage("Can't start because " + e.getMessage());
return false;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java
index f7405fd..7805bc5 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java
@@ -222,7 +222,8 @@ public class Import extends Configured implements Tool {
}
}
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while emitting Cell", e);
+ Thread.currentThread().interrupt();
}
}
@@ -286,7 +287,8 @@ public class Import extends Configured implements Tool {
}
}
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while emitting Cell", e);
+ Thread.currentThread().interrupt();
}
}
@@ -319,7 +321,8 @@ public class Import extends Configured implements Tool {
try {
writeResult(row, value, context);
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while writing result", e);
+ Thread.currentThread().interrupt();
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java
index 32b7561..a968870 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/SyncTable.java
@@ -779,7 +779,7 @@ public class SyncTable extends Configured implements Tool {
} catch (Exception e) {
- e.printStackTrace();
+ LOG.error("Failed to parse commandLine arguments", e);
printUsage("Can't start because " + e.getMessage());
return false;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java
index 8dc7156..5d40619 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java
@@ -39,14 +39,16 @@ import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Counter;
import org.apache.hadoop.mapreduce.Mapper;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* Write table content out to files in hdfs.
*/
@InterfaceAudience.Public
public class TsvImporterMapper
-extends Mapper<LongWritable, Text, ImmutableBytesWritable, Put>
-{
+ extends Mapper<LongWritable, Text, ImmutableBytesWritable, Put> {
+ private static final Logger LOG = LoggerFactory.getLogger(TsvImporterMapper.class);
/** Timestamp for all inserted rows */
protected long ts;
@@ -199,7 +201,8 @@ extends Mapper<LongWritable, Text, ImmutableBytesWritable, Put>
}
throw new IOException(badLine);
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while emitting put", e);
+ Thread.currentThread().interrupt();
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.java
index f3f81ec..0127f26 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.java
@@ -28,14 +28,16 @@ import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Counter;
import org.apache.yetus.audience.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* Write table content out to map output files.
*/
@InterfaceAudience.Public
public class TsvImporterTextMapper
-extends Mapper<LongWritable, Text, ImmutableBytesWritable, Text>
-{
+ extends Mapper<LongWritable, Text, ImmutableBytesWritable, Text> {
+ private static final Logger LOG = LoggerFactory.getLogger(TsvImporterTextMapper.class);
/** Column seperator */
private String separator;
@@ -121,7 +123,7 @@ extends Mapper<LongWritable, Text, ImmutableBytesWritable, Text>
}
throw new IOException(badLine);
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while emitting TSV text", e);
Thread.currentThread().interrupt();
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java
index fe8ff94..aa61316 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java
@@ -123,7 +123,8 @@ public class WALPlayer extends Configured implements Tool {
}
}
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while emitting Cell", e);
+ Thread.currentThread().interrupt();
}
}
@@ -199,7 +200,8 @@ public class WALPlayer extends Configured implements Tool {
}
}
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while writing results", e);
+ Thread.currentThread().interrupt();
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java
----------------------------------------------------------------------
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java
index 85eebc5..d1b5c60 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java
@@ -644,7 +644,7 @@ public class VerifyReplication extends Configured implements Tool {
}
} catch (Exception e) {
- e.printStackTrace();
+ LOG.error("Failed to parse commandLine arguments", e);
printUsage("Can't start because " + e.getMessage());
return false;
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
index c692365..b74efe4 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
@@ -32,6 +32,8 @@ import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
import org.apache.yetus.audience.InterfaceAudience;
import org.apache.yetus.audience.InterfaceStability;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hbase.procedure2.Procedure;
import org.apache.hadoop.hbase.procedure2.ProcedureUtil;
import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureWALEntry;
@@ -51,6 +53,8 @@ import org.apache.hbase.thirdparty.org.apache.commons.cli.PosixParser;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
@InterfaceStability.Evolving
public class ProcedureWALPrettyPrinter extends Configured implements Tool {
+ private static final Logger LOG = LoggerFactory.getLogger(ProcedureWALPrettyPrinter.class);
+
private final PrintStream out;
public ProcedureWALPrettyPrinter() {
@@ -171,7 +175,7 @@ public class ProcedureWALPrettyPrinter extends Configured implements Tool {
return(-1);
}
} catch (ParseException e) {
- e.printStackTrace();
+ LOG.error("Failed to parse commandLine arguments", e);
HelpFormatter formatter = new HelpFormatter();
formatter.printHelp("ProcedureWALPrettyPrinter ", options, true);
return(-1);
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
index 1914393..d3e62dc 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java
@@ -298,7 +298,8 @@ public class LocalHBaseCluster {
LOG.info("Waiting on " + rst.getRegionServer().toString());
rst.join();
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while waiting for {} to finish. Retrying join", rst.getName(), e);
+ Thread.currentThread().interrupt();
}
}
regionThreads.remove(rst);
@@ -370,7 +371,9 @@ public class LocalHBaseCluster {
LOG.info("Waiting on " + masterThread.getMaster().getServerName().toString());
masterThread.join();
} catch (InterruptedException e) {
- e.printStackTrace();
+ LOG.error("Interrupted while waiting for {} to finish. Retrying join",
+ masterThread.getName(), e);
+ Thread.currentThread().interrupt();
}
}
masterThreads.remove(masterThread);
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
index 1feeeeb..489894e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java
@@ -315,7 +315,7 @@ public class HMasterCommandLine extends ServerCommandLine {
try {
this.zkcluster.shutdown();
} catch (IOException e) {
- e.printStackTrace();
+ LOG.error("Failed to shutdown MiniZooKeeperCluster", e);
}
}
}
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALPrettyPrinter.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALPrettyPrinter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALPrettyPrinter.java
index 85877fe..3fa496f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALPrettyPrinter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALPrettyPrinter.java
@@ -50,6 +50,8 @@ import org.apache.hbase.thirdparty.org.apache.commons.cli.ParseException;
import org.apache.hbase.thirdparty.org.apache.commons.cli.PosixParser;
import com.fasterxml.jackson.databind.ObjectMapper;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* WALPrettyPrinter prints the contents of a given WAL with a variety of
@@ -67,6 +69,8 @@ import com.fasterxml.jackson.databind.ObjectMapper;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
@InterfaceStability.Evolving
public class WALPrettyPrinter {
+ private static final Logger LOG = LoggerFactory.getLogger(WALPrettyPrinter.class);
+
private boolean outputValues;
private boolean outputJSON;
// The following enable filtering by sequence, region, and row, respectively
@@ -400,7 +404,7 @@ public class WALPrettyPrinter {
if (cmd.hasOption("w"))
printer.setRowFilter(cmd.getOptionValue("w"));
} catch (ParseException e) {
- e.printStackTrace();
+ LOG.error("Failed to parse commandLine arguments", e);
HelpFormatter formatter = new HelpFormatter();
formatter.printHelp("HFile filename(s) ", options, true);
System.exit(-1);
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java
----------------------------------------------------------------------
diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java
index 3e75584..fc00327 100644
--- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java
+++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java
@@ -206,7 +206,7 @@ public class ThriftServer {
try {
this.infoServer.stop();
} catch (Exception ex) {
- ex.printStackTrace();
+ LOG.error("Failed to stop infoServer", ex);
}
}
serverRunner.shutdown();
http://git-wip-us.apache.org/repos/asf/hbase/blob/49ae8549/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
----------------------------------------------------------------------
diff --git a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
index 0193515..8e70611 100644
--- a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
+++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
@@ -44,6 +44,8 @@ import org.apache.zookeeper.server.ServerConfig;
import org.apache.zookeeper.server.ZooKeeperServerMain;
import org.apache.zookeeper.server.quorum.QuorumPeerConfig;
import org.apache.zookeeper.server.quorum.QuorumPeerMain;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* HBase's version of ZooKeeper's QuorumPeer. When HBase is set to manage
@@ -55,6 +57,8 @@ import org.apache.zookeeper.server.quorum.QuorumPeerMain;
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
@InterfaceStability.Evolving
public final class HQuorumPeer {
+ private static final Logger LOG = LoggerFactory.getLogger(HQuorumPeer.class);
+
private HQuorumPeer() {
}
@@ -77,7 +81,7 @@ public final class HQuorumPeer {
runZKServer(zkConfig);
} catch (Exception e) {
- e.printStackTrace();
+ LOG.error("Failed to start ZKServer", e);
System.exit(-1);
}
}