You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by su...@apache.org on 2012/04/07 01:03:55 UTC
svn commit: r1310631 -
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFsDefaultValue.java
Author: suresh
Date: Fri Apr 6 23:03:54 2012
New Revision: 1310631
URL: http://svn.apache.org/viewvc?rev=1310631&view=rev
Log:
HDFS-3121. Add HDFS tests for HADOOP-8014 change. Contributed by John George. Missed adding the file in the earlier commit.
Added:
hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFsDefaultValue.java
Added: hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFsDefaultValue.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFsDefaultValue.java?rev=1310631&view=auto
==============================================================================
--- hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFsDefaultValue.java (added)
+++ hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFsDefaultValue.java Fri Apr 6 23:03:54 2012
@@ -0,0 +1,176 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.fs.viewfs;
+
+
+import java.io.IOException;
+import java.net.URI;
+import java.net.URISyntaxException;
+
+import javax.security.auth.login.LoginException;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FileStatus;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.FileSystemTestHelper;
+import org.apache.hadoop.fs.FsConstants;
+import org.apache.hadoop.fs.ContentSummary;
+import org.apache.hadoop.hdfs.DistributedFileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
+import org.apache.hadoop.io.DataInputBuffer;
+import org.apache.hadoop.io.DataOutputBuffer;
+import org.apache.hadoop.security.UserGroupInformation;
+import org.apache.hadoop.fs.FsServerDefaults;
+import org.apache.hadoop.hdfs.DFSConfigKeys;
+import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_FILE_BUFFER_SIZE_DEFAULT;
+import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_FILE_BUFFER_SIZE_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_SIZE_DEFAULT;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_SIZE_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_DEFAULT;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_DEFAULT;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_SYNCONCLOSE_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_MIN_KEY;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_REPLICATION_DEFAULT;
+import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_REPLICATION_KEY;
+
+import org.junit.AfterClass;
+import org.junit.BeforeClass;
+import org.junit.Test;
+import static org.junit.Assert.*;
+
+/**
+ * Tests for viewfs implementation of default fs level values.
+ * This tests for both passing in a path (based on mount point)
+ * to obtain the default value of the fs that the path is mounted on
+ * or just passing in no arguments.
+ */
+public class TestViewFsDefaultValue {
+
+ static final String testFileDir = "/tmp/test/";
+ static final String testFileName = testFileDir + "testFileStatusSerialziation";
+ private static MiniDFSCluster cluster;
+ private static Configuration CONF = new Configuration();
+ private static FileSystem fHdfs;
+ private static FileSystem vfs;
+ private static Path testFilePath;
+ private static Path testFileDirPath;
+
+ @BeforeClass
+ public static void clusterSetupAtBegining() throws IOException,
+ LoginException, URISyntaxException {
+
+ CONF.setLong(DFS_BLOCK_SIZE_KEY, DFS_BLOCK_SIZE_DEFAULT);
+ CONF.setInt(DFS_BYTES_PER_CHECKSUM_KEY, DFS_BYTES_PER_CHECKSUM_DEFAULT);
+ CONF.setInt(DFS_CLIENT_WRITE_PACKET_SIZE_KEY,
+ DFS_CLIENT_WRITE_PACKET_SIZE_DEFAULT);
+ CONF.setInt(DFS_REPLICATION_KEY, DFS_REPLICATION_DEFAULT + 1);
+ CONF.setInt(IO_FILE_BUFFER_SIZE_KEY, IO_FILE_BUFFER_SIZE_DEFAULT);
+
+ cluster = new MiniDFSCluster.Builder(CONF).numDataNodes(DFS_REPLICATION_DEFAULT + 1).build();
+ cluster.waitClusterUp();
+ fHdfs = cluster.getFileSystem();
+ FileSystemTestHelper.createFile(fHdfs, testFileName);
+ Configuration conf = ViewFileSystemTestSetup.createConfig();
+ ConfigUtil.addLink(conf, "/tmp", new URI(fHdfs.getUri().toString() +
+ "/tmp"));
+ vfs = FileSystem.get(FsConstants.VIEWFS_URI, conf);
+ testFileDirPath = new Path (testFileDir);
+ testFilePath = new Path (testFileName);
+ }
+
+
+ /**
+ * Test that default blocksize values can be retrieved on the client side.
+ */
+ @Test
+ public void testGetDefaultBlockSize()
+ throws IOException, URISyntaxException {
+ // createFile does not use defaultBlockSize to create the file,
+ // but we are only looking at the defaultBlockSize, so this
+ // test should still pass
+ try {
+ vfs.getDefaultBlockSize();
+ fail("getServerDefaults on viewFs did not throw excetion!");
+ } catch (NotInMountpointException e) {
+ assertEquals(vfs.getDefaultBlockSize(testFilePath),
+ DFS_BLOCK_SIZE_DEFAULT);
+ }
+ }
+
+ /**
+ * Test that default replication values can be retrieved on the client side.
+ */
+ @Test
+ public void testGetDefaultReplication()
+ throws IOException, URISyntaxException {
+ try {
+ vfs.getDefaultReplication();
+ fail("getDefaultReplication on viewFs did not throw excetion!");
+ } catch (NotInMountpointException e) {
+ assertEquals(vfs.getDefaultReplication(testFilePath),
+ DFS_REPLICATION_DEFAULT+1);
+ }
+ }
+
+
+ /**
+ * Test that server default values can be retrieved on the client side.
+ */
+ @Test
+ public void testServerDefaults() throws IOException {
+ try {
+ FsServerDefaults serverDefaults = vfs.getServerDefaults();
+ fail("getServerDefaults on viewFs did not throw excetion!");
+ } catch (NotInMountpointException e) {
+ FsServerDefaults serverDefaults = vfs.getServerDefaults(testFilePath);
+ assertEquals(DFS_BLOCK_SIZE_DEFAULT, serverDefaults.getBlockSize());
+ assertEquals(DFS_BYTES_PER_CHECKSUM_DEFAULT,
+ serverDefaults.getBytesPerChecksum());
+ assertEquals(DFS_CLIENT_WRITE_PACKET_SIZE_DEFAULT,
+ serverDefaults.getWritePacketSize());
+ assertEquals(IO_FILE_BUFFER_SIZE_DEFAULT,
+ serverDefaults.getFileBufferSize());
+ assertEquals(DFS_REPLICATION_DEFAULT + 1,
+ serverDefaults.getReplication());
+ }
+ }
+
+ /**
+ * Test that getContentSummary can be retrieved on the client side.
+ */
+ @Test
+ public void testGetContentSummary() throws IOException {
+ FileSystem hFs = cluster.getFileSystem(0);
+ final DistributedFileSystem dfs = (DistributedFileSystem)hFs;
+ dfs.setQuota(testFileDirPath, 100, 500);
+ ContentSummary cs = vfs.getContentSummary(testFileDirPath);
+ assertEquals(100, cs.getQuota());
+ assertEquals(500, cs.getSpaceQuota());
+ }
+
+ @AfterClass
+ public static void cleanup() throws IOException {
+ fHdfs.delete(new Path(testFileName), true);
+ }
+
+}