You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by bu...@apache.org on 2016/02/13 18:19:10 UTC
hbase git commit: Revert "HBASE-9393 Hbase does not closing a closed
socket resulting in many CLOSE_WAIT"
Repository: hbase
Updated Branches:
refs/heads/master 141962888 -> 3897c4e10
Revert "HBASE-9393 Hbase does not closing a closed socket resulting in many CLOSE_WAIT"
This reverts commit c8d133186b85a4e7298bab3376ad96899860b1c7.
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/3897c4e1
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/3897c4e1
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/3897c4e1
Branch: refs/heads/master
Commit: 3897c4e102b223847e1ceedf363df59fe24eb294
Parents: 1419628
Author: Sean Busbey <bu...@cloudera.com>
Authored: Sat Feb 13 11:18:53 2016 -0600
Committer: Sean Busbey <bu...@cloudera.com>
Committed: Sat Feb 13 11:18:53 2016 -0600
----------------------------------------------------------------------
.../org/apache/hadoop/hbase/io/hfile/HFile.java | 34 ++++----------------
.../hadoop/hbase/io/hfile/HFileBlock.java | 20 +-----------
.../hadoop/hbase/io/hfile/HFileReaderImpl.java | 9 ------
3 files changed, 7 insertions(+), 56 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hbase/blob/3897c4e1/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
index 03d681e..1e1835f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
@@ -41,7 +41,6 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.CanUnbuffer;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileStatus;
@@ -475,11 +474,6 @@ public class HFile {
@VisibleForTesting
boolean prefetchComplete();
-
- /**
- * To close only the stream's socket. HBASE-9393
- */
- void unbufferStream();
}
/**
@@ -496,8 +490,8 @@ public class HFile {
*/
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="SF_SWITCH_FALLTHROUGH",
justification="Intentional")
- private static Reader openReader(Path path, FSDataInputStreamWrapper fsdis, long size,
- CacheConfig cacheConf, HFileSystem hfs, Configuration conf) throws IOException {
+ private static Reader pickReaderVersion(Path path, FSDataInputStreamWrapper fsdis,
+ long size, CacheConfig cacheConf, HFileSystem hfs, Configuration conf) throws IOException {
FixedFileTrailer trailer = null;
try {
boolean isHBaseChecksum = fsdis.shouldUseHBaseChecksum();
@@ -519,22 +513,6 @@ public class HFile {
LOG.warn("Error closing fsdis FSDataInputStreamWrapper", t2);
}
throw new CorruptHFileException("Problem reading HFile Trailer from file " + path, t);
- } finally {
- unbufferStream(fsdis);
- }
- }
-
- static void unbufferStream(FSDataInputStreamWrapper fsdis) {
- boolean useHBaseChecksum = fsdis.shouldUseHBaseChecksum();
- final FSDataInputStream stream = fsdis.getStream(useHBaseChecksum);
- if (stream != null && stream.getWrappedStream() instanceof CanUnbuffer) {
- // Enclosing unbuffer() in try-catch just to be on defensive side.
- try {
- stream.unbuffer();
- } catch (Throwable e) {
- LOG.error("Failed to unbuffer the stream so possibly there may be a TCP socket connection "
- + "left open in CLOSE_WAIT state.", e);
- }
}
}
@@ -563,7 +541,7 @@ public class HFile {
} else {
hfs = (HFileSystem)fs;
}
- return openReader(path, fsdis, size, cacheConf, hfs, conf);
+ return pickReaderVersion(path, fsdis, size, cacheConf, hfs, conf);
}
/**
@@ -578,8 +556,8 @@ public class HFile {
FileSystem fs, Path path, CacheConfig cacheConf, Configuration conf) throws IOException {
Preconditions.checkNotNull(cacheConf, "Cannot create Reader with null CacheConf");
FSDataInputStreamWrapper stream = new FSDataInputStreamWrapper(fs, path);
- return openReader(path, stream, fs.getFileStatus(path).getLen(), cacheConf, stream.getHfs(),
- conf);
+ return pickReaderVersion(path, stream, fs.getFileStatus(path).getLen(),
+ cacheConf, stream.getHfs(), conf);
}
/**
@@ -589,7 +567,7 @@ public class HFile {
FSDataInputStream fsdis, long size, CacheConfig cacheConf, Configuration conf)
throws IOException {
FSDataInputStreamWrapper wrapper = new FSDataInputStreamWrapper(fsdis);
- return openReader(path, wrapper, size, cacheConf, null, conf);
+ return pickReaderVersion(path, wrapper, size, cacheConf, null, conf);
}
/**
http://git-wip-us.apache.org/repos/asf/hbase/blob/3897c4e1/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
index 79b3e1c..e7a1e5e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
@@ -33,10 +33,10 @@ import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.classification.InterfaceAudience;
import org.apache.hadoop.hbase.fs.HFileSystem;
+import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
import org.apache.hadoop.hbase.io.ByteArrayOutputStream;
import org.apache.hadoop.hbase.io.ByteBuffInputStream;
import org.apache.hadoop.hbase.io.ByteBufferSupportDataOutputStream;
-import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
import org.apache.hadoop.hbase.io.encoding.HFileBlockDecodingContext;
import org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultDecodingContext;
@@ -1311,11 +1311,6 @@ public class HFileBlock implements Cacheable {
void setIncludesMemstoreTS(boolean includesMemstoreTS);
void setDataBlockEncoder(HFileDataBlockEncoder encoder);
-
- /**
- * To close only the stream's socket. HBASE-9393
- */
- void unbufferStream();
}
/**
@@ -1763,19 +1758,6 @@ public class HFileBlock implements Cacheable {
public String toString() {
return "hfs=" + hfs + ", path=" + pathName + ", fileContext=" + fileContext;
}
-
- @Override
- public void unbufferStream() {
- // To handle concurrent reads, ensure that no other client is accessing the streams while we
- // unbuffer it.
- if (streamLock.tryLock()) {
- try {
- HFile.unbufferStream(this.streamWrapper);
- } finally {
- streamLock.unlock();
- }
- }
- }
}
@Override
http://git-wip-us.apache.org/repos/asf/hbase/blob/3897c4e1/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
index f676e60..b2f5ded 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
@@ -575,10 +575,6 @@ public class HFileReaderImpl implements HFile.Reader, Configurable {
@Override
public void close() {
- if (!pread) {
- // For seek + pread stream socket should be closed when the scanner is closed. HBASE-9393
- reader.unbufferStream();
- }
this.returnBlocks(true);
}
@@ -1902,9 +1898,4 @@ public class HFileReaderImpl implements HFile.Reader, Configurable {
public int getMajorVersion() {
return 3;
}
-
- @Override
- public void unbufferStream() {
- fsBlockReader.unbufferStream();
- }
}