You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by ar...@apache.org on 2015/05/09 01:31:10 UTC
[1/2] hadoop git commit: HDFS-8097. TestFileTruncate is failing
intermittently. (Contributed by Rakesh R)
Repository: hadoop
Updated Branches:
refs/heads/branch-2 453488cff -> 969a8f6c9
refs/heads/trunk 565899884 -> 59995cec4
HDFS-8097. TestFileTruncate is failing intermittently. (Contributed by Rakesh R)
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/59995cec
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/59995cec
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/59995cec
Branch: refs/heads/trunk
Commit: 59995cec4ad9efcef7d4641375ca3eb40e2429ef
Parents: 5658998
Author: Arpit Agarwal <ar...@apache.org>
Authored: Fri May 8 16:27:52 2015 -0700
Committer: Arpit Agarwal <ar...@apache.org>
Committed: Fri May 8 16:28:24 2015 -0700
----------------------------------------------------------------------
hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 +++
.../hdfs/server/namenode/TestFileTruncate.java | 22 ++++++++------------
2 files changed, 12 insertions(+), 13 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/59995cec/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
index b766e26..0b721ee 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
+++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
@@ -723,6 +723,9 @@ Release 2.8.0 - UNRELEASED
HDFS-8326. Documentation about when checkpoints are run is out of date.
(Misty Stanley-Jones via xyao)
+ HDFS-8097. TestFileTruncate is failing intermittently. (Rakesh R via
+ Arpit Agarwal)
+
Release 2.7.1 - UNRELEASED
INCOMPATIBLE CHANGES
http://git-wip-us.apache.org/repos/asf/hadoop/blob/59995cec/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
index bd19a5b..1f65f78 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
@@ -63,6 +63,7 @@ import org.apache.hadoop.util.Time;
import org.apache.hadoop.util.ToolRunner;
import org.apache.log4j.Level;
import org.junit.AfterClass;
+import org.junit.Before;
import org.junit.BeforeClass;
import org.junit.Test;
@@ -87,6 +88,8 @@ public class TestFileTruncate {
static MiniDFSCluster cluster;
static DistributedFileSystem fs;
+ private Path parent;
+
@BeforeClass
public static void startUp() throws IOException {
conf = new HdfsConfiguration();
@@ -110,6 +113,12 @@ public class TestFileTruncate {
if(cluster != null) cluster.shutdown();
}
+ @Before
+ public void setup() throws IOException {
+ parent = new Path("/test");
+ fs.delete(parent, true);
+ }
+
/**
* Truncate files of different sizes byte by byte.
*/
@@ -117,7 +126,6 @@ public class TestFileTruncate {
public void testBasicTruncate() throws IOException {
int startingFileSize = 3 * BLOCK_SIZE;
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
@@ -257,7 +265,6 @@ public class TestFileTruncate {
*/
void testSnapshotWithAppendTruncate(int ... deleteOrder) throws IOException {
FSDirectory fsDir = cluster.getNamesystem().getFSDirectory();
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
fs.allowSnapshot(parent);
@@ -421,7 +428,6 @@ public class TestFileTruncate {
}
void testSnapshotWithTruncates(int ... deleteOrder) throws IOException {
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
fs.allowSnapshot(parent);
@@ -664,7 +670,6 @@ public class TestFileTruncate {
public void testTruncateWithDataNodesRestart() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testTruncateWithDataNodesRestart");
writeContents(contents, startingFileSize, p);
@@ -719,7 +724,6 @@ public class TestFileTruncate {
public void testCopyOnTruncateWithDataNodesRestart() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testCopyOnTruncateWithDataNodesRestart");
writeContents(contents, startingFileSize, p);
@@ -779,7 +783,6 @@ public class TestFileTruncate {
public void testTruncateWithDataNodesRestartImmediately() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testTruncateWithDataNodesRestartImmediately");
writeContents(contents, startingFileSize, p);
@@ -839,7 +842,6 @@ public class TestFileTruncate {
public void testTruncateWithDataNodesShutdownImmediately() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testTruncateWithDataNodesShutdownImmediately");
writeContents(contents, startingFileSize, p);
@@ -905,7 +907,6 @@ public class TestFileTruncate {
*/
@Test
public void testUpgradeAndRestart() throws IOException {
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
fs.allowSnapshot(parent);
@@ -994,7 +995,6 @@ public class TestFileTruncate {
FSNamesystem fsn = cluster.getNamesystem();
String client = "client";
String clientMachine = "clientMachine";
- Path parent = new Path("/test");
String src = "/test/testTruncateRecovery";
Path srcPath = new Path(src);
@@ -1067,7 +1067,6 @@ public class TestFileTruncate {
@Test
public void testTruncateShellCommand() throws Exception {
- final Path parent = new Path("/test");
final Path src = new Path("/test/testTruncateShellCommand");
final int oldLength = 2*BLOCK_SIZE + 1;
final int newLength = BLOCK_SIZE + 1;
@@ -1084,7 +1083,6 @@ public class TestFileTruncate {
@Test
public void testTruncateShellCommandOnBlockBoundary() throws Exception {
- final Path parent = new Path("/test");
final Path src = new Path("/test/testTruncateShellCommandOnBoundary");
final int oldLength = 2 * BLOCK_SIZE;
final int newLength = BLOCK_SIZE;
@@ -1100,7 +1098,6 @@ public class TestFileTruncate {
@Test
public void testTruncateShellCommandWithWaitOption() throws Exception {
- final Path parent = new Path("/test");
final Path src = new Path("/test/testTruncateShellCommandWithWaitOption");
final int oldLength = 2 * BLOCK_SIZE + 1;
final int newLength = BLOCK_SIZE + 1;
@@ -1136,7 +1133,6 @@ public class TestFileTruncate {
public void testTruncate4Symlink() throws IOException {
final int fileLength = 3 * BLOCK_SIZE;
- final Path parent = new Path("/test");
fs.mkdirs(parent);
final byte[] contents = AppendTestUtil.initBuffer(fileLength);
final Path file = new Path(parent, "testTruncate4Symlink");
[2/2] hadoop git commit: HDFS-8097. TestFileTruncate is failing
intermittently. (Contributed by Rakesh R)
Posted by ar...@apache.org.
HDFS-8097. TestFileTruncate is failing intermittently. (Contributed by Rakesh R)
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/969a8f6c
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/969a8f6c
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/969a8f6c
Branch: refs/heads/branch-2
Commit: 969a8f6c96a6e710723e3b460157b8064e65e3e7
Parents: 453488c
Author: Arpit Agarwal <ar...@apache.org>
Authored: Fri May 8 16:27:52 2015 -0700
Committer: Arpit Agarwal <ar...@apache.org>
Committed: Fri May 8 16:29:01 2015 -0700
----------------------------------------------------------------------
hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 6 ++++++
.../hdfs/server/namenode/TestFileTruncate.java | 22 ++++++++------------
2 files changed, 15 insertions(+), 13 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/969a8f6c/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
index 00d0ea6..2c5686e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
+++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
@@ -393,6 +393,12 @@ Release 2.8.0 - UNRELEASED
HDFS-8311. DataStreamer.transfer() should timeout the socket InputStream.
(Esteban Gutierrez via Yongjun Zhang)
+ HDFS-8326. Documentation about when checkpoints are run is out of date.
+ (Misty Stanley-Jones via xyao)
+
+ HDFS-8097. TestFileTruncate is failing intermittently. (Rakesh R via
+ Arpit Agarwal)
+
Release 2.7.1 - UNRELEASED
INCOMPATIBLE CHANGES
http://git-wip-us.apache.org/repos/asf/hadoop/blob/969a8f6c/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
index bd19a5b..1f65f78 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java
@@ -63,6 +63,7 @@ import org.apache.hadoop.util.Time;
import org.apache.hadoop.util.ToolRunner;
import org.apache.log4j.Level;
import org.junit.AfterClass;
+import org.junit.Before;
import org.junit.BeforeClass;
import org.junit.Test;
@@ -87,6 +88,8 @@ public class TestFileTruncate {
static MiniDFSCluster cluster;
static DistributedFileSystem fs;
+ private Path parent;
+
@BeforeClass
public static void startUp() throws IOException {
conf = new HdfsConfiguration();
@@ -110,6 +113,12 @@ public class TestFileTruncate {
if(cluster != null) cluster.shutdown();
}
+ @Before
+ public void setup() throws IOException {
+ parent = new Path("/test");
+ fs.delete(parent, true);
+ }
+
/**
* Truncate files of different sizes byte by byte.
*/
@@ -117,7 +126,6 @@ public class TestFileTruncate {
public void testBasicTruncate() throws IOException {
int startingFileSize = 3 * BLOCK_SIZE;
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
@@ -257,7 +265,6 @@ public class TestFileTruncate {
*/
void testSnapshotWithAppendTruncate(int ... deleteOrder) throws IOException {
FSDirectory fsDir = cluster.getNamesystem().getFSDirectory();
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
fs.allowSnapshot(parent);
@@ -421,7 +428,6 @@ public class TestFileTruncate {
}
void testSnapshotWithTruncates(int ... deleteOrder) throws IOException {
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
fs.allowSnapshot(parent);
@@ -664,7 +670,6 @@ public class TestFileTruncate {
public void testTruncateWithDataNodesRestart() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testTruncateWithDataNodesRestart");
writeContents(contents, startingFileSize, p);
@@ -719,7 +724,6 @@ public class TestFileTruncate {
public void testCopyOnTruncateWithDataNodesRestart() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testCopyOnTruncateWithDataNodesRestart");
writeContents(contents, startingFileSize, p);
@@ -779,7 +783,6 @@ public class TestFileTruncate {
public void testTruncateWithDataNodesRestartImmediately() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testTruncateWithDataNodesRestartImmediately");
writeContents(contents, startingFileSize, p);
@@ -839,7 +842,6 @@ public class TestFileTruncate {
public void testTruncateWithDataNodesShutdownImmediately() throws Exception {
int startingFileSize = 3 * BLOCK_SIZE;
byte[] contents = AppendTestUtil.initBuffer(startingFileSize);
- final Path parent = new Path("/test");
final Path p = new Path(parent, "testTruncateWithDataNodesShutdownImmediately");
writeContents(contents, startingFileSize, p);
@@ -905,7 +907,6 @@ public class TestFileTruncate {
*/
@Test
public void testUpgradeAndRestart() throws IOException {
- Path parent = new Path("/test");
fs.mkdirs(parent);
fs.setQuota(parent, 100, 1000);
fs.allowSnapshot(parent);
@@ -994,7 +995,6 @@ public class TestFileTruncate {
FSNamesystem fsn = cluster.getNamesystem();
String client = "client";
String clientMachine = "clientMachine";
- Path parent = new Path("/test");
String src = "/test/testTruncateRecovery";
Path srcPath = new Path(src);
@@ -1067,7 +1067,6 @@ public class TestFileTruncate {
@Test
public void testTruncateShellCommand() throws Exception {
- final Path parent = new Path("/test");
final Path src = new Path("/test/testTruncateShellCommand");
final int oldLength = 2*BLOCK_SIZE + 1;
final int newLength = BLOCK_SIZE + 1;
@@ -1084,7 +1083,6 @@ public class TestFileTruncate {
@Test
public void testTruncateShellCommandOnBlockBoundary() throws Exception {
- final Path parent = new Path("/test");
final Path src = new Path("/test/testTruncateShellCommandOnBoundary");
final int oldLength = 2 * BLOCK_SIZE;
final int newLength = BLOCK_SIZE;
@@ -1100,7 +1098,6 @@ public class TestFileTruncate {
@Test
public void testTruncateShellCommandWithWaitOption() throws Exception {
- final Path parent = new Path("/test");
final Path src = new Path("/test/testTruncateShellCommandWithWaitOption");
final int oldLength = 2 * BLOCK_SIZE + 1;
final int newLength = BLOCK_SIZE + 1;
@@ -1136,7 +1133,6 @@ public class TestFileTruncate {
public void testTruncate4Symlink() throws IOException {
final int fileLength = 3 * BLOCK_SIZE;
- final Path parent = new Path("/test");
fs.mkdirs(parent);
final byte[] contents = AppendTestUtil.initBuffer(fileLength);
final Path file = new Path(parent, "testTruncate4Symlink");