You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2015/10/28 16:45:55 UTC
[07/14] hive git commit: HIVE-12237 : Use slf4j as logging facade
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java
index b7b6f90..a1e35cb 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java
@@ -28,8 +28,8 @@ import java.util.Random;
import java.util.regex.Matcher;
import java.util.regex.Pattern;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.QueryPlan;
import org.apache.hadoop.hive.ql.exec.Task;
@@ -51,7 +51,7 @@ public class HiveHistoryImpl implements HiveHistory{
String histFileName; // History file name
- private static final Log LOG = LogFactory.getLog("hive.ql.exec.HiveHistoryImpl");
+ private static final Logger LOG = LoggerFactory.getLogger("hive.ql.exec.HiveHistoryImpl");
private static final Random randGen = new Random();
@@ -355,7 +355,7 @@ public class HiveHistoryImpl implements HiveHistory{
@Override
public void closeStream() {
- IOUtils.cleanup(LOG, histStream);
+ IOUtils.closeStream(histStream);
}
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryViewer.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryViewer.java b/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryViewer.java
index 1b357de..616f2d6 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryViewer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryViewer.java
@@ -22,8 +22,8 @@ import java.io.IOException;
import java.util.HashMap;
import java.util.Map;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.history.HiveHistory.Keys;
import org.apache.hadoop.hive.ql.history.HiveHistory.Listener;
import org.apache.hadoop.hive.ql.history.HiveHistory.QueryInfo;
@@ -38,7 +38,7 @@ public class HiveHistoryViewer implements Listener {
String historyFile;
String sessionId;
- private static final Log LOG = LogFactory.getLog(HiveHistoryViewer.class);
+ private static final Logger LOG = LoggerFactory.getLogger(HiveHistoryViewer.class);
// Job Hash Map
private final HashMap<String, QueryInfo> jobInfoMap = new HashMap<String, QueryInfo>();
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java b/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java
index 2caa7ae..5610fab 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java
@@ -22,8 +22,8 @@ import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.TimeUnit;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.QueryPlan;
@@ -46,7 +46,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder;
*/
public class ATSHook implements ExecuteWithHookContext {
- private static final Log LOG = LogFactory.getLog(ATSHook.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(ATSHook.class.getName());
private static final Object LOCK = new Object();
private static ExecutorService executor;
private static TimelineClient timelineClient;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java b/ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java
index 9988c79..64220f2 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/hooks/LineageLogger.java
@@ -31,8 +31,8 @@ import java.util.Set;
import org.apache.commons.collections.SetUtils;
import org.apache.commons.io.output.StringBuilderWriter;
import org.apache.commons.lang.StringUtils;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.ObjectPair;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.metastore.api.FieldSchema;
@@ -61,7 +61,7 @@ import com.google.gson.stream.JsonWriter;
*/
public class LineageLogger implements ExecuteWithHookContext {
- private static final Log LOG = LogFactory.getLog(LineageLogger.class);
+ private static final Logger LOG = LoggerFactory.getLogger(LineageLogger.class);
private static final HashSet<String> OPERATION_NAMES = new HashSet<String>();
@@ -178,10 +178,10 @@ public class LineageLogger implements ExecuteWithHookContext {
writer.endObject();
writer.close();
- // Log the lineage info
+ // Logger the lineage info
String lineage = out.toString();
if (testMode) {
- // Log to console
+ // Logger to console
log(lineage);
} else {
// In non-test mode, emit to a log file,
@@ -199,7 +199,7 @@ public class LineageLogger implements ExecuteWithHookContext {
}
/**
- * Log an error to console if available.
+ * Logger an error to console if available.
*/
private void log(String error) {
LogHelper console = SessionState.getConsole();
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecOrcFileDump.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecOrcFileDump.java b/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecOrcFileDump.java
index b0b4a36..d5d1370 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecOrcFileDump.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecOrcFileDump.java
@@ -21,8 +21,8 @@ import java.io.IOException;
import java.io.PrintStream;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -45,7 +45,7 @@ import com.google.common.collect.Lists;
* in the file just to verify the impact of bloom filter fpp.
*/
public class PostExecOrcFileDump implements ExecuteWithHookContext {
- private static final Log LOG = LogFactory.getLog(PostExecOrcFileDump.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(PostExecOrcFileDump.class.getName());
private static final PathFilter hiddenFileFilter = new PathFilter() {
public boolean accept(Path p) {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecTezSummaryPrinter.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecTezSummaryPrinter.java b/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecTezSummaryPrinter.java
index 60c587f..81bda08 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecTezSummaryPrinter.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/hooks/PostExecTezSummaryPrinter.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.hive.ql.hooks;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.QueryPlan;
import org.apache.hadoop.hive.ql.exec.Utilities;
@@ -34,7 +34,7 @@ import org.apache.tez.common.counters.TezCounters;
* Post execution hook to print hive tez counters to console error stream.
*/
public class PostExecTezSummaryPrinter implements ExecuteWithHookContext {
- private static final Log LOG = LogFactory.getLog(PostExecTezSummaryPrinter.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(PostExecTezSummaryPrinter.class.getName());
@Override
public void run(HookContext hookContext) throws Exception {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/hooks/WriteEntity.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/hooks/WriteEntity.java b/ql/src/java/org/apache/hadoop/hive/ql/hooks/WriteEntity.java
index 298e7f0..515f8b2 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/hooks/WriteEntity.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/hooks/WriteEntity.java
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.hooks;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.metastore.api.Database;
import org.apache.hadoop.hive.ql.metadata.DummyPartition;
@@ -35,7 +35,7 @@ import java.io.Serializable;
*/
public class WriteEntity extends Entity implements Serializable {
- private static final Log LOG = LogFactory.getLog(WriteEntity.class);
+ private static final Logger LOG = LoggerFactory.getLogger(WriteEntity.class);
private boolean isTempURI = false;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java b/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java
index 835caf1..36bc9cd 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java
@@ -17,15 +17,15 @@
*/
package org.apache.hadoop.hive.ql.index;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* Holds index related constants
*/
public class HiveIndex {
- public static final Log l4j = LogFactory.getLog("HiveIndex");
+ public static final Logger l4j = LoggerFactory.getLogger("HiveIndex");
public static String INDEX_TABLE_CREATETIME = "hive.index.basetbl.dfs.lastModifiedTime";
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexResult.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexResult.java b/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexResult.java
index 6fe200b..33cc5c3 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexResult.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexResult.java
@@ -25,8 +25,8 @@ import java.util.Map;
import java.util.SortedSet;
import java.util.TreeSet;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
@@ -47,8 +47,8 @@ import org.apache.hadoop.mapred.LineRecordReader.LineReader;
*/
public class HiveIndexResult {
- public static final Log l4j =
- LogFactory.getLog(HiveIndexResult.class.getSimpleName());
+ public static final Logger l4j =
+ LoggerFactory.getLogger(HiveIndexResult.class.getSimpleName());
// IndexBucket
static class IBucket {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexedInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexedInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexedInputFormat.java
index c62add0..e072ee6 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexedInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndexedInputFormat.java
@@ -27,8 +27,8 @@ import java.util.Map;
import java.util.Arrays;
import java.util.HashMap;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
@@ -52,7 +52,7 @@ import org.apache.hadoop.mapred.JobConf;
* Uses a blockfilter file to specify the blocks to query.
*/
public class HiveIndexedInputFormat extends HiveInputFormat {
- public static final Log l4j = LogFactory.getLog("HiveIndexInputFormat");
+ public static final Logger l4j = LoggerFactory.getLogger("HiveIndexInputFormat");
private final String indexFile;
public HiveIndexedInputFormat() {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/index/bitmap/BitmapIndexHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/index/bitmap/BitmapIndexHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/index/bitmap/BitmapIndexHandler.java
index cb191ac..5ddbd0b 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/index/bitmap/BitmapIndexHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/index/bitmap/BitmapIndexHandler.java
@@ -25,8 +25,8 @@ import java.util.List;
import java.util.Map;
import java.util.Set;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.common.StatsSetupConst;
import org.apache.hadoop.hive.conf.HiveConf;
@@ -64,7 +64,7 @@ import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPLessThan;
public class BitmapIndexHandler extends TableBasedIndexHandler {
private Configuration configuration;
- private static final Log LOG = LogFactory.getLog(BitmapIndexHandler.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(BitmapIndexHandler.class.getName());
@Override
public void generateIndexQuery(List<Index> indexes, ExprNodeDesc predicate,
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/index/compact/CompactIndexHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/index/compact/CompactIndexHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/index/compact/CompactIndexHandler.java
index 586e16d..1d9e131 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/index/compact/CompactIndexHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/index/compact/CompactIndexHandler.java
@@ -25,8 +25,8 @@ import java.util.LinkedHashMap;
import java.util.List;
import java.util.Set;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.common.JavaUtils;
import org.apache.hadoop.hive.conf.HiveConf;
@@ -73,7 +73,7 @@ public class CompactIndexHandler extends TableBasedIndexHandler {
private Set<String> partitionCols;
// Whether or not the conditions have been met to use the fact the index is sorted
private boolean useSorted;
- private static final Log LOG = LogFactory.getLog(CompactIndexHandler.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(CompactIndexHandler.class.getName());
@Override
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/index/compact/HiveCompactIndexInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/index/compact/HiveCompactIndexInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/index/compact/HiveCompactIndexInputFormat.java
index 7cebe68..b71084d 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/index/compact/HiveCompactIndexInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/index/compact/HiveCompactIndexInputFormat.java
@@ -18,14 +18,14 @@
package org.apache.hadoop.hive.ql.index.compact;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.index.HiveIndexedInputFormat;
public class HiveCompactIndexInputFormat extends HiveIndexedInputFormat {
- public static final Log l4j =
- LogFactory.getLog(HiveCompactIndexInputFormat.class.getSimpleName());
+ public static final Logger l4j =
+ LoggerFactory.getLogger(HiveCompactIndexInputFormat.class.getSimpleName());
public HiveCompactIndexInputFormat() {
super("hive.index.compact.file");
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java
index e8d070c..24137e2 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.io;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
@@ -91,7 +91,7 @@ public class AcidUtils {
private AcidUtils() {
// NOT USED
}
- private static final Log LOG = LogFactory.getLog(AcidUtils.class);
+ private static final Logger LOG = LoggerFactory.getLogger(AcidUtils.class);
private static final Pattern ORIGINAL_PATTERN =
Pattern.compile("[0-9]+_[0-9]+");
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/BucketizedHiveInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/BucketizedHiveInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/BucketizedHiveInputFormat.java
index edcc3b6..a9c1614 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/BucketizedHiveInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/BucketizedHiveInputFormat.java
@@ -22,8 +22,8 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -50,8 +50,8 @@ import org.apache.hadoop.mapred.Reporter;
public class BucketizedHiveInputFormat<K extends WritableComparable, V extends Writable>
extends HiveInputFormat<K, V> {
- public static final Log LOG = LogFactory
- .getLog("org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat");
+ public static final Logger LOG = LoggerFactory
+ .getLogger("org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat");
@Override
public RecordReader getRecordReader(InputSplit split, JobConf job,
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/CodecPool.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/CodecPool.java b/ql/src/java/org/apache/hadoop/hive/ql/io/CodecPool.java
index 9dfb6bf..56b3f84 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/CodecPool.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/CodecPool.java
@@ -22,8 +22,8 @@ import java.util.HashMap;
import java.util.List;
import java.util.Map;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.io.compress.CompressionCodec;
import org.apache.hadoop.io.compress.Compressor;
import org.apache.hadoop.io.compress.Decompressor;
@@ -33,7 +33,7 @@ import org.apache.hadoop.io.compress.Decompressor;
* native) compression/decompression codecs.
*/
public final class CodecPool {
- private static final Log LOG = LogFactory.getLog(CodecPool.class);
+ private static final Logger LOG = LoggerFactory.getLogger(CodecPool.class);
/**
* A global compressor pool used to save the expensive
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java
index 53bc1fa..323ac43 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java
@@ -33,8 +33,8 @@ import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.Future;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -72,7 +72,7 @@ public class CombineHiveInputFormat<K extends WritableComparable, V extends Writ
extends HiveInputFormat<K, V> {
private static final String CLASS_NAME = CombineHiveInputFormat.class.getName();
- public static final Log LOG = LogFactory.getLog(CLASS_NAME);
+ public static final Logger LOG = LoggerFactory.getLogger(CLASS_NAME);
// max number of threads we can use to check non-combinable paths
private static final int MAX_CHECK_NONCOMBINABLE_THREAD_NUM = 50;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/HdfsUtils.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/HdfsUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/HdfsUtils.java
index 0095d31..38c99fd 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/HdfsUtils.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/HdfsUtils.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.io;
import java.io.IOException;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hdfs.DistributedFileSystem;
@@ -30,7 +30,7 @@ import org.apache.hadoop.hive.shims.ShimLoader;
public class HdfsUtils {
private static final HadoopShims SHIMS = ShimLoader.getHadoopShims();
- private static final Log LOG = LogFactory.getLog(HdfsUtils.class);
+ private static final Logger LOG = LoggerFactory.getLogger(HdfsUtils.class);
public static long getFileId(FileSystem fileSystem, Path path) throws IOException {
String pathStr = path.toUri().getPath();
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java
index 738ca9c..7d36e42 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java
@@ -23,8 +23,8 @@ import java.util.ArrayList;
import java.util.List;
import java.util.Map;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil;
@@ -56,7 +56,7 @@ import org.apache.hadoop.mapred.RecordReader;
*/
public abstract class HiveContextAwareRecordReader<K, V> implements RecordReader<K, V> {
- private static final Log LOG = LogFactory.getLog(HiveContextAwareRecordReader.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(HiveContextAwareRecordReader.class.getName());
private boolean initDone = false;
private long rangeStart;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java
index 06d3df7..0328a23 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java
@@ -33,8 +33,8 @@ import java.util.Properties;
import java.util.Set;
import java.util.concurrent.ConcurrentHashMap;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
@@ -77,7 +77,7 @@ import org.apache.hive.common.util.ReflectionUtil;
*
*/
public final class HiveFileFormatUtils {
- private static final Log LOG = LogFactory.getLog(HiveFileFormatUtils.class);
+ private static final Logger LOG = LoggerFactory.getLogger(HiveFileFormatUtils.class);
static {
outputFormatSubstituteMap =
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java
index 45ee9c5..29c4b61 100755
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java
@@ -30,8 +30,8 @@ import java.util.Set;
import java.util.concurrent.ConcurrentHashMap;
import java.util.Map.Entry;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configurable;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
@@ -77,7 +77,7 @@ public class HiveInputFormat<K extends WritableComparable, V extends Writable>
implements InputFormat<K, V>, JobConfigurable {
private static final String CLASS_NAME = HiveInputFormat.class.getName();
- private static final Log LOG = LogFactory.getLog(CLASS_NAME);
+ private static final Logger LOG = LoggerFactory.getLogger(CLASS_NAME);
/**
* A cache of InputFormat instances.
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/IOContextMap.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/IOContextMap.java b/ql/src/java/org/apache/hadoop/hive/ql/io/IOContextMap.java
index 14656a7..e857cf9 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/IOContextMap.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/IOContextMap.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.io;
import java.util.concurrent.ConcurrentHashMap;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.exec.Utilities;
@@ -44,7 +44,7 @@ import org.apache.hadoop.hive.ql.exec.Utilities;
*/
public class IOContextMap {
public static final String DEFAULT_CONTEXT = "";
- private static final Log LOG = LogFactory.getLog(IOContextMap.class);
+ private static final Logger LOG = LoggerFactory.getLogger(IOContextMap.class);
/** Used for Tez and MR */
private static final ConcurrentHashMap<String, IOContext> globalMap =
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/NullRowsInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/NullRowsInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/NullRowsInputFormat.java
index fd60fed..9638f2a 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/NullRowsInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/NullRowsInputFormat.java
@@ -22,8 +22,8 @@ import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.mapred.InputFormat;
import org.apache.hadoop.mapred.InputSplit;
@@ -39,7 +39,7 @@ public class NullRowsInputFormat implements InputFormat<NullWritable, NullWritab
JobConfigurable {
static final int MAX_ROW = 100; // to prevent infinite loop
- static final Log LOG = LogFactory.getLog(NullRowsRecordReader.class.getName());
+ static final Logger LOG = LoggerFactory.getLogger(NullRowsRecordReader.class.getName());
public static class DummyInputSplit implements InputSplit {
public DummyInputSplit() {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java b/ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java
index 2a27676..d391164 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java
@@ -29,8 +29,8 @@ import java.security.MessageDigest;
import java.util.Arrays;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.ChecksumException;
import org.apache.hadoop.fs.FSDataInputStream;
@@ -340,7 +340,7 @@ import static org.apache.hadoop.hive.conf.HiveConf.ConfVars.*;
*/
public class RCFile {
- private static final Log LOG = LogFactory.getLog(RCFile.class);
+ private static final Logger LOG = LoggerFactory.getLogger(RCFile.class);
// internal variable
public static final String COLUMN_NUMBER_METADATA_STR = "hive.io.rcfile.column.number";
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/StorageFormatFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/StorageFormatFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/io/StorageFormatFactory.java
index e0bf153..12d9d9a 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/StorageFormatFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/StorageFormatFactory.java
@@ -24,13 +24,13 @@ import java.util.ServiceLoader;
import javax.annotation.Nullable;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import com.google.common.collect.ImmutableMap;
public class StorageFormatFactory {
- private static final Log LOG = LogFactory.getLog(StorageFormatFactory.class);
+ private static final Logger LOG = LoggerFactory.getLogger(StorageFormatFactory.class);
private final Map<String, StorageFormatDescriptor> storageFormats;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroContainerOutputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroContainerOutputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroContainerOutputFormat.java
index 2784185..59d3bba 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroContainerOutputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroContainerOutputFormat.java
@@ -30,8 +30,8 @@ import org.apache.avro.file.CodecFactory;
import org.apache.avro.file.DataFileWriter;
import org.apache.avro.generic.GenericDatumWriter;
import org.apache.avro.generic.GenericRecord;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.exec.FileSinkOperator;
@@ -52,7 +52,7 @@ import org.apache.hadoop.util.Progressable;
public class AvroContainerOutputFormat
implements HiveOutputFormat<WritableComparable, AvroGenericRecordWritable> {
- public static final Log LOG = LogFactory.getLog(AvroContainerOutputFormat.class);
+ public static final Logger LOG = LoggerFactory.getLogger(AvroContainerOutputFormat.class);
@Override
public org.apache.hadoop.hive.ql.exec.FileSinkOperator.RecordWriter getHiveRecordWriter(JobConf jobConf,
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroGenericRecordReader.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroGenericRecordReader.java b/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroGenericRecordReader.java
index 89fac3f..30862c8 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroGenericRecordReader.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/avro/AvroGenericRecordReader.java
@@ -29,8 +29,8 @@ import org.apache.avro.generic.GenericData;
import org.apache.avro.generic.GenericDatumReader;
import org.apache.avro.generic.GenericRecord;
import org.apache.avro.mapred.FsInput;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.exec.Utilities;
import org.apache.hadoop.hive.ql.plan.MapWork;
@@ -52,7 +52,7 @@ import org.apache.hadoop.mapred.Reporter;
*/
public class AvroGenericRecordReader implements
RecordReader<NullWritable, AvroGenericRecordWritable>, JobConfigurable {
- private static final Log LOG = LogFactory.getLog(AvroGenericRecordReader.class);
+ private static final Logger LOG = LoggerFactory.getLogger(AvroGenericRecordReader.class);
final private org.apache.avro.file.FileReader<GenericRecord> reader;
final private long start;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileMapper.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileMapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileMapper.java
index 4c5eed2..1816fac 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileMapper.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileMapper.java
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.io.merge;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.exec.AbstractFileMergeOperator;
import org.apache.hadoop.hive.ql.exec.Operator;
import org.apache.hadoop.hive.ql.exec.Utilities;
@@ -40,7 +40,7 @@ import java.io.IOException;
* name is used for serialization and deserialization of MergeFileWork.
*/
public class MergeFileMapper extends MapReduceBase implements Mapper {
- public static final Log LOG = LogFactory.getLog("MergeFileMapper");
+ public static final Logger LOG = LoggerFactory.getLogger("MergeFileMapper");
private static final String PLAN_KEY = "__MAP_PLAN__";
private JobConf jc;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java b/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java
index bd50b46..2f09014 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileTask.java
@@ -189,7 +189,7 @@ public class MergeFileTask extends Task<MergeFileWork> implements Serializable,
}
} catch (Exception e) {
// jobClose needs to execute successfully otherwise fail task
- LOG.warn(e);
+ LOG.warn("Job close failed ",e);
if (success) {
success = false;
returnVal = 3;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileWork.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileWork.java b/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileWork.java
index fbc87e8..3339c8d 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileWork.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/merge/MergeFileWork.java
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.io.merge;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -46,7 +46,7 @@ import java.util.List;
@Explain(displayName = "Merge File Operator", explainLevels = { Level.USER, Level.DEFAULT, Level.EXTENDED })
public class MergeFileWork extends MapWork {
- private static final Log LOG = LogFactory.getLog(MergeFileWork.class);
+ private static final Logger LOG = LoggerFactory.getLogger(MergeFileWork.class);
private List<Path> inputPaths;
private Path outputDir;
private boolean hasDynamicPartitions;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java
index 3dde0c4..2275188 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java
@@ -24,8 +24,8 @@ import java.util.ArrayList;
import java.util.List;
import java.util.ListIterator;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.io.DiskRange;
import org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.BufferChunk;
@@ -34,7 +34,7 @@ import com.google.protobuf.CodedInputStream;
public abstract class InStream extends InputStream {
- private static final Log LOG = LogFactory.getLog(InStream.class);
+ private static final Logger LOG = LoggerFactory.getLogger(InStream.class);
private static final int PROTOBUF_MESSAGE_MAX_LIMIT = 1024 << 20; // 1GB
protected final Long fileId;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/MemoryManager.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/MemoryManager.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/MemoryManager.java
index 0347a1c..4d5f735 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/MemoryManager.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/MemoryManager.java
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.io.orc;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
@@ -43,7 +43,7 @@ import java.util.concurrent.locks.ReentrantLock;
*/
class MemoryManager {
- private static final Log LOG = LogFactory.getLog(MemoryManager.class);
+ private static final Logger LOG = LoggerFactory.getLogger(MemoryManager.class);
/**
* How often should we check the memory sizes? Measured in rows added
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java
index 7c927dc..73037ea 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java
@@ -38,8 +38,8 @@ import java.util.concurrent.Future;
import java.util.concurrent.atomic.AtomicInteger;
import org.apache.commons.codec.binary.Hex;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.BlockLocation;
import org.apache.hadoop.fs.FileStatus;
@@ -123,7 +123,7 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
ETL
}
- private static final Log LOG = LogFactory.getLog(OrcInputFormat.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcInputFormat.class);
private static boolean isDebugEnabled = LOG.isDebugEnabled();
static final HadoopShims SHIMS = ShimLoader.getHadoopShims();
static final String MIN_SPLIT_SIZE =
@@ -466,7 +466,7 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
private final boolean cacheStripeDetails;
private final AtomicInteger cacheHitCounter = new AtomicInteger(0);
private final AtomicInteger numFilesCounter = new AtomicInteger(0);
- private ValidTxnList transactionList;
+ private final ValidTxnList transactionList;
private SplitStrategyKind splitStrategyKind;
private final SearchArgument sarg;
@@ -851,7 +851,7 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
private final boolean hasBase;
private OrcFile.WriterVersion writerVersion;
private long projColsUncompressedSize;
- private List<OrcSplit> deltaSplits;
+ private final List<OrcSplit> deltaSplits;
public SplitGenerator(SplitInfo splitInfo) throws IOException {
this.context = splitInfo.context;
@@ -1140,7 +1140,7 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
context, adi.fs, adi.splitPath, adi.acidInfo, adi.baseOrOriginalFiles);
if (isDebugEnabled) {
- LOG.debug(splitStrategy);
+ LOG.debug("Split strategy: ", splitStrategy);
}
// Hack note - different split strategies return differently typed lists, yay Java.
@@ -1218,10 +1218,10 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
private final Long fileId;
private final List<StripeInformation> stripeInfos;
private FileMetaInfo fileMetaInfo;
- private List<StripeStatistics> stripeStats;
- private List<OrcProto.ColumnStatistics> fileStats;
- private List<OrcProto.Type> types;
- private OrcFile.WriterVersion writerVersion;
+ private final List<StripeStatistics> stripeStats;
+ private final List<OrcProto.ColumnStatistics> fileStats;
+ private final List<OrcProto.Type> types;
+ private final OrcFile.WriterVersion writerVersion;
FileInfo(long modificationTime, long size, List<StripeInformation> stripeInfos,
@@ -1296,6 +1296,7 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
private final RecordIdentifier id;
private final RowReader<OrcStruct> inner;
+ @Override
public RecordIdentifier getRecordIdentifier() {
return id;
}
@@ -1606,7 +1607,7 @@ public class OrcInputFormat implements InputFormat<NullWritable, OrcStruct>,
/** Local footer cache using Guava. Stores convoluted Java objects. */
private static class LocalCache implements FooterCache {
- private Cache<Path, FileInfo> cache;
+ private final Cache<Path, FileInfo> cache;
public LocalCache(int numThreads, int cacheStripeDetailsSize) {
cache = CacheBuilder.newBuilder()
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcNewInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcNewInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcNewInputFormat.java
index 1833d3d..c15b35f 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcNewInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcNewInputFormat.java
@@ -21,8 +21,8 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.shims.ShimLoader;
@@ -38,7 +38,7 @@ import org.apache.hadoop.mapreduce.lib.input.FileSplit;
* value is the OrcStruct object */
public class OrcNewInputFormat extends InputFormat<NullWritable, OrcStruct>{
- private static final Log LOG = LogFactory.getLog(OrcNewInputFormat.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcNewInputFormat.class);
@Override
public RecordReader<NullWritable, OrcStruct> createRecordReader(
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcOutputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcOutputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcOutputFormat.java
index ad24c58..8a5de7f 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcOutputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcOutputFormat.java
@@ -24,8 +24,8 @@ import java.util.Arrays;
import java.util.List;
import java.util.Properties;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.io.AcidOutputFormat;
@@ -66,7 +66,7 @@ import org.apache.hadoop.util.Progressable;
public class OrcOutputFormat extends FileOutputFormat<NullWritable, OrcSerdeRow>
implements AcidOutputFormat<NullWritable, OrcSerdeRow> {
- private static final Log LOG = LogFactory.getLog(OrcOutputFormat.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcOutputFormat.class);
static TypeDescription convertTypeInfo(TypeInfo info) {
switch (info.getCategory()) {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRawRecordMerger.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRawRecordMerger.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRawRecordMerger.java
index fb5110d..ebe1afd 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRawRecordMerger.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRawRecordMerger.java
@@ -18,8 +18,8 @@
package org.apache.hadoop.hive.ql.io.orc;
import com.google.common.annotations.VisibleForTesting;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FileSystem;
@@ -52,7 +52,7 @@ import java.util.TreeMap;
*/
public class OrcRawRecordMerger implements AcidInputFormat.RawReader<OrcStruct>{
- private static final Log LOG = LogFactory.getLog(OrcRawRecordMerger.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcRawRecordMerger.class);
private final Configuration conf;
private final boolean collapse;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRecordUpdater.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRecordUpdater.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRecordUpdater.java
index 01374a7..67c5a11 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRecordUpdater.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcRecordUpdater.java
@@ -25,8 +25,8 @@ import java.nio.charset.CharsetDecoder;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
@@ -51,7 +51,7 @@ import com.google.common.annotations.VisibleForTesting;
*/
public class OrcRecordUpdater implements RecordUpdater {
- private static final Log LOG = LogFactory.getLog(OrcRecordUpdater.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcRecordUpdater.class);
public static final String ACID_KEY_INDEX_NAME = "hive.acid.key.index";
public static final String ACID_FORMAT = "_orc_acid_version";
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSerde.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSerde.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSerde.java
index 595f3b3..c0e9b1a 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSerde.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSerde.java
@@ -23,8 +23,8 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.Properties;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedSerde;
@@ -46,7 +46,7 @@ import org.apache.hadoop.io.Writable;
@SerDeSpec(schemaProps = {serdeConstants.LIST_COLUMNS, serdeConstants.LIST_COLUMN_TYPES, OrcSerde.COMPRESSION})
public class OrcSerde implements SerDe, VectorizedSerde {
- private static final Log LOG = LogFactory.getLog(OrcSerde.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcSerde.class);
private final OrcSerdeRow row = new OrcSerdeRow();
private ObjectInspector inspector = null;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSplit.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSplit.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSplit.java
index 33513bc..81afb48 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSplit.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcSplit.java
@@ -25,8 +25,8 @@ import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.io.ColumnarSplit;
import org.apache.hadoop.hive.ql.io.AcidInputFormat;
@@ -42,7 +42,7 @@ import org.apache.hadoop.mapred.FileSplit;
*
*/
public class OrcSplit extends FileSplit implements ColumnarSplit {
- private static final Log LOG = LogFactory.getLog(OrcSplit.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcSplit.class);
private FileMetaInfo fileMetaInfo;
private boolean hasFooter;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcUtils.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcUtils.java
index 3e2af23..b654b64 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcUtils.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcUtils.java
@@ -20,11 +20,11 @@ package org.apache.hadoop.hive.ql.io.orc;
import java.util.Arrays;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
public class OrcUtils {
- private static final Log LOG = LogFactory.getLog(OrcUtils.class);
+ private static final Logger LOG = LoggerFactory.getLogger(OrcUtils.class);
/**
* Returns selected columns as a boolean array with true value set for specified column names.
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java
index 15c4417..f6dea25 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java
@@ -26,8 +26,8 @@ import java.util.Collections;
import java.util.List;
import java.util.Set;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FileSystem;
@@ -49,7 +49,7 @@ import com.google.protobuf.CodedInputStream;
public class ReaderImpl implements Reader {
- private static final Log LOG = LogFactory.getLog(ReaderImpl.class);
+ private static final Logger LOG = LoggerFactory.getLogger(ReaderImpl.class);
private static final int DIRECTORY_SIZE_GUESS = 16 * 1024;
@@ -278,7 +278,7 @@ public class ReaderImpl implements Reader {
* @param path the data source path for error messages
* @param version the version of hive that wrote the file.
*/
- static void checkOrcVersion(Log log, Path path, List<Integer> version) {
+ static void checkOrcVersion(Logger log, Path path, List<Integer> version) {
if (version.size() >= 1) {
int major = version.get(0);
int minor = 0;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderFactory.java
index 23a9af4..5e7d636 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderFactory.java
@@ -21,8 +21,8 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.serde.serdeConstants;
import org.apache.hadoop.hive.serde2.objectinspector.ListObjectInspector;
@@ -46,7 +46,7 @@ import com.google.common.collect.Lists;
* to see if type promotions are possible.
*/
public class RecordReaderFactory {
- static final Log LOG = LogFactory.getLog(RecordReaderFactory.class);
+ static final Logger LOG = LoggerFactory.getLogger(RecordReaderFactory.class);
private static final boolean isLogInfoEnabled = LOG.isInfoEnabled();
public static TreeReaderFactory.TreeReader createTreeReader(int colId,
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java
index 33c914e..0696277 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java
@@ -29,8 +29,8 @@ import java.util.List;
import java.util.Map;
import org.apache.commons.lang3.exception.ExceptionUtils;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -51,7 +51,7 @@ import org.apache.hadoop.hive.serde2.io.TimestampWritable;
import org.apache.hadoop.io.Text;
public class RecordReaderImpl implements RecordReader {
- static final Log LOG = LogFactory.getLog(RecordReaderImpl.class);
+ static final Logger LOG = LoggerFactory.getLogger(RecordReaderImpl.class);
private static final boolean isLogDebugEnabled = LOG.isDebugEnabled();
private final Path path;
private final long firstRow;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java
index 8318a6a..aef3231 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java
@@ -21,8 +21,8 @@ import java.io.EOFException;
import java.io.IOException;
import java.util.Arrays;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.ErrorMsg;
import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
import org.apache.hadoop.hive.ql.io.orc.RunLengthIntegerWriterV2.EncodingType;
@@ -33,7 +33,7 @@ import org.apache.hadoop.hive.ql.io.orc.RunLengthIntegerWriterV2.EncodingType;
* compression techniques.
*/
public class RunLengthIntegerReaderV2 implements IntegerReader {
- public static final Log LOG = LogFactory.getLog(RunLengthIntegerReaderV2.class);
+ public static final Logger LOG = LoggerFactory.getLogger(RunLengthIntegerReaderV2.class);
private InStream input;
private final boolean signed;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java
index 5bcd8f4..5a82d20 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java
@@ -32,8 +32,8 @@ import java.util.Map;
import java.util.TimeZone;
import java.util.TreeMap;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
@@ -102,7 +102,7 @@ import com.google.protobuf.CodedOutputStream;
*/
public class WriterImpl implements Writer, MemoryManager.Callback {
- private static final Log LOG = LogFactory.getLog(WriterImpl.class);
+ private static final Logger LOG = LoggerFactory.getLogger(WriterImpl.class);
static final HadoopShims SHIMS = ShimLoader.getHadoopShims();
private static final int HDFS_BUFFER_SIZE = 256 * 1024;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java
index 0d3b64c..e0c0743 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java
@@ -22,8 +22,8 @@ import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.Pool;
import org.apache.hadoop.hive.common.Pool.PoolObjectHelper;
import org.apache.hadoop.hive.common.io.DataCache;
@@ -83,7 +83,7 @@ import org.apache.hadoop.hive.ql.io.orc.encoded.Reader.PoolFactory;
* not use it; thus, at the end we go thru all the MBs, and release those not released by (5).
*/
class EncodedReaderImpl implements EncodedReader {
- public static final Log LOG = LogFactory.getLog(EncodedReaderImpl.class);
+ public static final Logger LOG = LoggerFactory.getLogger(EncodedReaderImpl.class);
private static final Object POOLS_CREATION_LOCK = new Object();
private static Pools POOLS;
private static class Pools {
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java
index 3e00612..f95ebcd 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java
@@ -13,8 +13,8 @@
*/
package org.apache.hadoop.hive.ql.io.parquet;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.io.sarg.PredicateLeaf;
import org.apache.hadoop.hive.ql.io.sarg.PredicateLeaf.Operator;
@@ -35,7 +35,7 @@ import static org.apache.parquet.filter2.predicate.FilterApi.floatColumn;
import static org.apache.parquet.filter2.predicate.FilterApi.intColumn;
public class LeafFilterFactory {
- private static final Log LOG = LogFactory.getLog(LeafFilterFactory.class);
+ private static final Logger LOG = LoggerFactory.getLogger(LeafFilterFactory.class);
class IntFilterPredicateLeafBuilder extends FilterPredicateLeafBuilder {
/**
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java
index d82e93c..a4e35cb 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java
@@ -14,8 +14,8 @@
package org.apache.hadoop.hive.ql.io.parquet;
import java.io.IOException;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.exec.Utilities;
import org.apache.hadoop.hive.ql.io.parquet.read.DataWritableReadSupport;
import org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper;
@@ -36,7 +36,7 @@ import org.apache.parquet.hadoop.ParquetInputFormat;
*/
public class MapredParquetInputFormat extends FileInputFormat<NullWritable, ArrayWritable> {
- private static final Log LOG = LogFactory.getLog(MapredParquetInputFormat.class);
+ private static final Logger LOG = LoggerFactory.getLogger(MapredParquetInputFormat.class);
private final ParquetInputFormat<ArrayWritable> realInput;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java
index 5e71df9..bfb48a9 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java
@@ -19,8 +19,8 @@ import java.util.Arrays;
import java.util.List;
import java.util.Properties;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.ql.io.HiveOutputFormat;
@@ -50,7 +50,7 @@ import org.apache.parquet.hadoop.ParquetOutputFormat;
public class MapredParquetOutputFormat extends FileOutputFormat<NullWritable, ParquetHiveRecord>
implements HiveOutputFormat<NullWritable, ParquetHiveRecord> {
- private static final Log LOG = LogFactory.getLog(MapredParquetOutputFormat.class);
+ private static final Logger LOG = LoggerFactory.getLogger(MapredParquetOutputFormat.class);
protected ParquetOutputFormat<ParquetHiveRecord> realOutputFormat;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/ProjectionPusher.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/ProjectionPusher.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/ProjectionPusher.java
index 4848efd..13390de 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/ProjectionPusher.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/ProjectionPusher.java
@@ -22,8 +22,8 @@ import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.exec.Operator;
@@ -38,7 +38,7 @@ import org.apache.hadoop.mapred.JobConf;
public class ProjectionPusher {
- private static final Log LOG = LogFactory.getLog(ProjectionPusher.class);
+ private static final Logger LOG = LoggerFactory.getLogger(ProjectionPusher.class);
private final Map<String, PartitionDesc> pathToPartitionInfo =
new LinkedHashMap<String, PartitionDesc>();
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java
index ed99615..b28d870 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java
@@ -14,8 +14,8 @@
package org.apache.hadoop.hive.ql.io.parquet;
import java.io.IOException;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.exec.vector.VectorColumnAssign;
import org.apache.hadoop.hive.ql.exec.vector.VectorColumnAssignFactory;
import org.apache.hadoop.hive.ql.exec.vector.VectorizedInputFormatInterface;
@@ -41,14 +41,14 @@ import org.apache.parquet.hadoop.ParquetInputFormat;
public class VectorizedParquetInputFormat extends FileInputFormat<NullWritable, VectorizedRowBatch>
implements VectorizedInputFormatInterface {
- private static final Log LOG = LogFactory.getLog(VectorizedParquetInputFormat.class);
+ private static final Logger LOG = LoggerFactory.getLogger(VectorizedParquetInputFormat.class);
/**
* Vectorized record reader for vectorized Parquet input format
*/
private static class VectorizedParquetRecordReader implements
RecordReader<NullWritable, VectorizedRowBatch> {
- private static final Log LOG = LogFactory.getLog(VectorizedParquetRecordReader.class);
+ private static final Logger LOG = LoggerFactory.getLogger(VectorizedParquetRecordReader.class);
private final ParquetRecordReaderWrapper internalReader;
private VectorizedRowBatchCtx rbCtx;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetFilterPredicateConverter.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetFilterPredicateConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetFilterPredicateConverter.java
index d1864ae..786a260 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetFilterPredicateConverter.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetFilterPredicateConverter.java
@@ -17,8 +17,8 @@
*/
package org.apache.hadoop.hive.ql.io.parquet.read;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.ql.io.parquet.FilterPredicateLeafBuilder;
import org.apache.hadoop.hive.ql.io.parquet.LeafFilterFactory;
import org.apache.hadoop.hive.ql.io.sarg.ExpressionTree;
@@ -34,7 +34,7 @@ import java.util.List;
import java.util.Set;
public class ParquetFilterPredicateConverter {
- private static final Log LOG = LogFactory.getLog(ParquetFilterPredicateConverter.class);
+ private static final Logger LOG = LoggerFactory.getLogger(ParquetFilterPredicateConverter.class);
/**
* Translate the search argument to the filter predicate parquet uses. It includes
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java
index f689b90..74a1a82 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java
@@ -17,8 +17,8 @@ import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.conf.HiveConf;
@@ -55,7 +55,7 @@ import org.apache.parquet.schema.MessageTypeParser;
import com.google.common.base.Strings;
public class ParquetRecordReaderWrapper implements RecordReader<NullWritable, ArrayWritable> {
- public static final Log LOG = LogFactory.getLog(ParquetRecordReaderWrapper.class);
+ public static final Logger LOG = LoggerFactory.getLogger(ParquetRecordReaderWrapper.class);
private final long splitLen; // for getPos()
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java
index 493cd36..69272dc 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java
@@ -13,8 +13,8 @@
*/
package org.apache.hadoop.hive.ql.io.parquet.write;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe;
import org.apache.hadoop.hive.ql.io.parquet.timestamp.NanoTimeUtils;
@@ -59,7 +59,7 @@ import java.util.Map;
* This class is only used through DataWritableWriteSupport class.
*/
public class DataWritableWriter {
- private static final Log LOG = LogFactory.getLog(DataWritableWriter.class);
+ private static final Logger LOG = LoggerFactory.getLogger(DataWritableWriter.class);
protected final RecordConsumer recordConsumer;
private final GroupType schema;
@@ -547,4 +547,4 @@ public class DataWritableWriter {
recordConsumer.addInteger(DateWritable.dateToDays(vDate));
}
}
-}
\ No newline at end of file
+}
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java
index 9e2a9e1..2f838fc 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java
@@ -16,8 +16,8 @@ package org.apache.hadoop.hive.ql.io.parquet.write;
import java.io.IOException;
import java.util.Properties;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.NullWritable;
@@ -39,7 +39,7 @@ import org.apache.parquet.hadoop.util.ContextUtil;
public class ParquetRecordWriterWrapper implements RecordWriter<NullWritable, ParquetHiveRecord>,
org.apache.hadoop.hive.ql.exec.FileSinkOperator.RecordWriter {
- public static final Log LOG = LogFactory.getLog(ParquetRecordWriterWrapper.class);
+ public static final Logger LOG = LoggerFactory.getLogger(ParquetRecordWriterWrapper.class);
private final org.apache.hadoop.mapreduce.RecordWriter<NullWritable, ParquetHiveRecord> realWriter;
private final TaskAttemptContext taskContext;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanMapper.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanMapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanMapper.java
index d06f502..8a5360e 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanMapper.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanMapper.java
@@ -22,8 +22,8 @@ import java.io.IOException;
import java.util.HashMap;
import java.util.Map;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.StatsSetupConst;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.ErrorMsg;
@@ -62,7 +62,7 @@ public class PartialScanMapper extends MapReduceBase implements
private boolean exception = false;
private Reporter rp = null;
- public final static Log LOG = LogFactory.getLog("PartialScanMapper");
+ private static final Logger LOG = LoggerFactory.getLogger("PartialScanMapper");
public PartialScanMapper() {
}
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java
index 8bebd0f..fd04fb5 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java
@@ -24,8 +24,7 @@ import java.util.ArrayList;
import java.util.List;
import org.apache.commons.lang.StringUtils;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -253,7 +252,7 @@ public class PartialScanTask extends Task<PartialScanWork> implements
jobID = rj.getID().toString();
}
} catch (Exception e) {
- LOG.warn(e);
+ LOG.warn("Failed in cleaning up ", e);
} finally {
HadoopJobExecHelper.runningJobs.remove(rj);
}
@@ -333,7 +332,7 @@ public class PartialScanTask extends Task<PartialScanWork> implements
}
HiveConf hiveConf = new HiveConf(conf, PartialScanTask.class);
- Log LOG = LogFactory.getLog(PartialScanTask.class.getName());
+ org.slf4j.Logger LOG = LoggerFactory.getLogger(PartialScanTask.class.getName());
boolean isSilent = HiveConf.getBoolVar(conf,
HiveConf.ConfVars.HIVESESSIONSILENT);
LogHelper console = new LogHelper(LOG, isSilent);
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateMapper.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateMapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateMapper.java
index 6d8694b..34a18cb 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateMapper.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateMapper.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hive.ql.io.rcfile.truncate;
import java.io.IOException;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
@@ -68,7 +68,7 @@ public class ColumnTruncateMapper extends MapReduceBase implements
Path dpPath;
ColumnTruncateWork work;
- public final static Log LOG = LogFactory.getLog(ColumnTruncateMapper.class.getName());
+ private static final Logger LOG = LoggerFactory.getLogger(ColumnTruncateMapper.class.getName());
public ColumnTruncateMapper() {
}
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java
index 149ad93..79b3cfa 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/truncate/ColumnTruncateTask.java
@@ -43,7 +43,6 @@ import org.apache.hadoop.hive.ql.session.SessionState;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.mapred.Counters;
import org.apache.hadoop.mapred.FileInputFormat;
-import org.apache.hadoop.mapred.InputFormat;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.RunningJob;
@@ -121,7 +120,7 @@ public class ColumnTruncateTask extends Task<ColumnTruncateWork> implements Seri
LOG.info("Using " + inpFormat);
try {
- job.setInputFormat((Class<? extends InputFormat>) JavaUtils.loadClass(inpFormat));
+ job.setInputFormat(JavaUtils.loadClass(inpFormat));
} catch (ClassNotFoundException e) {
throw new RuntimeException(e.getMessage(), e);
}
@@ -218,7 +217,7 @@ public class ColumnTruncateTask extends Task<ColumnTruncateWork> implements Seri
ColumnTruncateMapper.jobClose(outputPath, success, job, console,
work.getDynPartCtx(), null);
} catch (Exception e) {
- LOG.warn(e);
+ LOG.warn("Failed while cleaning up ", e);
} finally {
HadoopJobExecHelper.runningJobs.remove(rj);
}
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/ConvertAstToSearchArg.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/ConvertAstToSearchArg.java b/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/ConvertAstToSearchArg.java
index 0fff9aa..7e888bc 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/ConvertAstToSearchArg.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/ConvertAstToSearchArg.java
@@ -23,8 +23,8 @@ import java.sql.Timestamp;
import java.util.List;
import org.apache.commons.codec.binary.Base64;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hive.common.type.HiveChar;
import org.apache.hadoop.hive.ql.exec.Utilities;
@@ -56,7 +56,7 @@ import com.esotericsoftware.kryo.Kryo;
import com.esotericsoftware.kryo.io.Input;
public class ConvertAstToSearchArg {
- private static final Log LOG = LogFactory.getLog(ConvertAstToSearchArg.class);
+ private static final Logger LOG = LoggerFactory.getLogger(ConvertAstToSearchArg.class);
private final SearchArgument.Builder builder =
SearchArgumentFactory.newBuilder();
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbLockManager.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbLockManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbLockManager.java
index bb9da9d..42616ac 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbLockManager.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbLockManager.java
@@ -17,8 +17,8 @@
*/
package org.apache.hadoop.hive.ql.lockmgr;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.JavaUtils;
import org.apache.hadoop.hive.common.metrics.common.Metrics;
import org.apache.hadoop.hive.common.metrics.common.MetricsConstant;
@@ -42,7 +42,7 @@ import java.util.Set;
public class DbLockManager implements HiveLockManager{
static final private String CLASS_NAME = DbLockManager.class.getName();
- static final private Log LOG = LogFactory.getLog(CLASS_NAME);
+ static final private Logger LOG = LoggerFactory.getLogger(CLASS_NAME);
private static final long MAX_SLEEP = 15000;
private HiveLockManagerCtx context;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java
index 219a54a..97d2282 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DbTxnManager.java
@@ -17,8 +17,8 @@
*/
package org.apache.hadoop.hive.ql.lockmgr;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.JavaUtils;
import org.apache.hadoop.hive.common.ValidTxnList;
import org.apache.hadoop.hive.conf.HiveConf;
@@ -47,7 +47,7 @@ import java.util.List;
public class DbTxnManager extends HiveTxnManagerImpl {
static final private String CLASS_NAME = DbTxnManager.class.getName();
- static final private Log LOG = LogFactory.getLog(CLASS_NAME);
+ static final private Logger LOG = LoggerFactory.getLogger(CLASS_NAME);
private DbLockManager lockMgr = null;
private IMetaStoreClient client = null;
http://git-wip-us.apache.org/repos/asf/hive/blob/55337444/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java
index 7acc53f..2d30198 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/DummyTxnManager.java
@@ -17,8 +17,8 @@
*/
package org.apache.hadoop.hive.ql.lockmgr;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.common.ValidTxnList;
import org.apache.hadoop.hive.common.ValidReadTxnList;
import org.apache.hadoop.hive.conf.HiveConf;
@@ -43,8 +43,8 @@ import java.util.*;
* transactions. This provides default Hive behavior.
*/
class DummyTxnManager extends HiveTxnManagerImpl {
- static final private Log LOG =
- LogFactory.getLog(DummyTxnManager.class.getName());
+ static final private Logger LOG =
+ LoggerFactory.getLogger(DummyTxnManager.class.getName());
private HiveLockManager lockMgr;