You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by wa...@apache.org on 2013/09/20 01:42:18 UTC
svn commit: r1524865 [3/3] - in
/hadoop/common/branches/HDFS-4949/hadoop-common-project: hadoop-auth/
hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/
hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/
hadoo...
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java Thu Sep 19 23:42:10 2013
@@ -216,13 +216,13 @@ public class TestIPC {
}
}
- @Test
+ @Test(timeout=60000)
public void testSerial() throws IOException, InterruptedException {
- testSerial(3, false, 2, 5, 100);
- testSerial(3, true, 2, 5, 10);
+ internalTestSerial(3, false, 2, 5, 100);
+ internalTestSerial(3, true, 2, 5, 10);
}
- public void testSerial(int handlerCount, boolean handlerSleep,
+ public void internalTestSerial(int handlerCount, boolean handlerSleep,
int clientCount, int callerCount, int callCount)
throws IOException, InterruptedException {
Server server = new TestServer(handlerCount, handlerSleep);
@@ -249,7 +249,7 @@ public class TestIPC {
server.stop();
}
- @Test
+ @Test(timeout=60000)
public void testStandAloneClient() throws IOException {
Client client = new Client(LongWritable.class, conf);
InetSocketAddress address = new InetSocketAddress("127.0.0.1", 10);
@@ -383,7 +383,7 @@ public class TestIPC {
}
}
- @Test
+ @Test(timeout=60000)
public void testIOEOnClientWriteParam() throws Exception {
doErrorTest(IOEOnWriteWritable.class,
LongWritable.class,
@@ -391,7 +391,7 @@ public class TestIPC {
LongWritable.class);
}
- @Test
+ @Test(timeout=60000)
public void testRTEOnClientWriteParam() throws Exception {
doErrorTest(RTEOnWriteWritable.class,
LongWritable.class,
@@ -399,7 +399,7 @@ public class TestIPC {
LongWritable.class);
}
- @Test
+ @Test(timeout=60000)
public void testIOEOnServerReadParam() throws Exception {
doErrorTest(LongWritable.class,
IOEOnReadWritable.class,
@@ -407,7 +407,7 @@ public class TestIPC {
LongWritable.class);
}
- @Test
+ @Test(timeout=60000)
public void testRTEOnServerReadParam() throws Exception {
doErrorTest(LongWritable.class,
RTEOnReadWritable.class,
@@ -416,7 +416,7 @@ public class TestIPC {
}
- @Test
+ @Test(timeout=60000)
public void testIOEOnServerWriteResponse() throws Exception {
doErrorTest(LongWritable.class,
LongWritable.class,
@@ -424,7 +424,7 @@ public class TestIPC {
LongWritable.class);
}
- @Test
+ @Test(timeout=60000)
public void testRTEOnServerWriteResponse() throws Exception {
doErrorTest(LongWritable.class,
LongWritable.class,
@@ -432,7 +432,7 @@ public class TestIPC {
LongWritable.class);
}
- @Test
+ @Test(timeout=60000)
public void testIOEOnClientReadResponse() throws Exception {
doErrorTest(LongWritable.class,
LongWritable.class,
@@ -440,7 +440,7 @@ public class TestIPC {
IOEOnReadWritable.class);
}
- @Test
+ @Test(timeout=60000)
public void testRTEOnClientReadResponse() throws Exception {
doErrorTest(LongWritable.class,
LongWritable.class,
@@ -453,7 +453,7 @@ public class TestIPC {
* that a ping should have been sent. This is a reproducer for a
* deadlock seen in one iteration of HADOOP-6762.
*/
- @Test
+ @Test(timeout=60000)
public void testIOEOnWriteAfterPingClient() throws Exception {
// start server
Client.setPingInterval(conf, 100);
@@ -481,7 +481,7 @@ public class TestIPC {
* Test that, if the socket factory throws an IOE, it properly propagates
* to the client.
*/
- @Test
+ @Test(timeout=60000)
public void testSocketFactoryException() throws IOException {
SocketFactory mockFactory = mock(SocketFactory.class);
doThrow(new IOException("Injected fault")).when(mockFactory).createSocket();
@@ -503,7 +503,7 @@ public class TestIPC {
* failure is handled properly. This is a regression test for
* HADOOP-7428.
*/
- @Test
+ @Test(timeout=60000)
public void testRTEDuringConnectionSetup() throws IOException {
// Set up a socket factory which returns sockets which
// throw an RTE when setSoTimeout is called.
@@ -544,7 +544,7 @@ public class TestIPC {
}
}
- @Test
+ @Test(timeout=60000)
public void testIpcTimeout() throws IOException {
// start server
Server server = new TestServer(1, true);
@@ -566,7 +566,7 @@ public class TestIPC {
addr, null, null, 3*PING_INTERVAL+MIN_SLEEP_TIME, conf);
}
- @Test
+ @Test(timeout=60000)
public void testIpcConnectTimeout() throws IOException {
// start server
Server server = new TestServer(1, true);
@@ -670,31 +670,31 @@ public class TestIPC {
return FD_DIR.list().length;
}
- @Test
+ @Test(timeout=60000)
public void testIpcFromHadoop_0_18_13() throws IOException {
doIpcVersionTest(NetworkTraces.HADOOP_0_18_3_RPC_DUMP,
NetworkTraces.RESPONSE_TO_HADOOP_0_18_3_RPC);
}
- @Test
+ @Test(timeout=60000)
public void testIpcFromHadoop0_20_3() throws IOException {
doIpcVersionTest(NetworkTraces.HADOOP_0_20_3_RPC_DUMP,
NetworkTraces.RESPONSE_TO_HADOOP_0_20_3_RPC);
}
- @Test
+ @Test(timeout=60000)
public void testIpcFromHadoop0_21_0() throws IOException {
doIpcVersionTest(NetworkTraces.HADOOP_0_21_0_RPC_DUMP,
NetworkTraces.RESPONSE_TO_HADOOP_0_21_0_RPC);
}
- @Test
+ @Test(timeout=60000)
public void testHttpGetResponse() throws IOException {
doIpcVersionTest("GET / HTTP/1.0\r\n\r\n".getBytes(),
Server.RECEIVED_HTTP_REQ_RESPONSE.getBytes());
}
- @Test
+ @Test(timeout=60000)
public void testConnectionRetriesOnSocketTimeoutExceptions() throws IOException {
Configuration conf = new Configuration();
// set max retries to 0
@@ -720,7 +720,7 @@ public class TestIPC {
* (1) the rpc server uses the call id/retry provided by the rpc client, and
* (2) the rpc client receives the same call id/retry from the rpc server.
*/
- @Test
+ @Test(timeout=60000)
public void testCallIdAndRetry() throws IOException {
final CallInfo info = new CallInfo();
@@ -772,7 +772,7 @@ public class TestIPC {
/**
* Test the retry count while used in a retry proxy.
*/
- @Test
+ @Test(timeout=60000)
public void testRetryProxy() throws IOException {
final Client client = new Client(LongWritable.class, conf);
@@ -785,7 +785,9 @@ public class TestIPC {
}
};
- final int totalRetry = 256;
+ // try more times, so it is easier to find race condition bug
+ // 10000 times runs about 6s on a core i7 machine
+ final int totalRetry = 10000;
DummyProtocol proxy = (DummyProtocol) Proxy.newProxyInstance(
DummyProtocol.class.getClassLoader(),
new Class[] { DummyProtocol.class }, new TestInvocationHandler(client,
@@ -807,7 +809,7 @@ public class TestIPC {
/**
* Test if the rpc server gets the default retry count (0) from client.
*/
- @Test
+ @Test(timeout=60000)
public void testInitialCallRetryCount() throws IOException {
// Override client to store the call id
final Client client = new Client(LongWritable.class, conf);
@@ -838,7 +840,7 @@ public class TestIPC {
/**
* Test if the rpc server gets the retry count from client.
*/
- @Test
+ @Test(timeout=60000)
public void testCallRetryCount() throws IOException {
final int retryCount = 255;
// Override client to store the call id
@@ -873,7 +875,7 @@ public class TestIPC {
* even if multiple threads are using the same client.
* @throws InterruptedException
*/
- @Test
+ @Test(timeout=60000)
public void testUniqueSequentialCallIds()
throws IOException, InterruptedException {
int serverThreads = 10, callerCount = 100, perCallerCallCount = 100;
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/log/TestLogLevel.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/log/TestLogLevel.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/log/TestLogLevel.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/log/TestLogLevel.java Thu Sep 19 23:42:10 2013
@@ -42,7 +42,10 @@ public class TestLogLevel extends TestCa
log.error("log.error1");
assertTrue(!Level.ERROR.equals(log.getEffectiveLevel()));
- HttpServer server = new HttpServer("..", "localhost", 22222, true);
+ HttpServer server = new HttpServer.Builder().setName("..")
+ .setBindAddress("localhost").setPort(22222).setFindPort(true)
+ .build();
+
server.start();
int port = server.getPort();
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestDoAsEffectiveUser.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestDoAsEffectiveUser.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestDoAsEffectiveUser.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestDoAsEffectiveUser.java Thu Sep 19 23:42:10 2013
@@ -38,6 +38,7 @@ import org.apache.hadoop.security.UserGr
import org.apache.hadoop.security.authorize.ProxyUsers;
import org.apache.hadoop.security.token.Token;
import org.apache.hadoop.security.token.TokenInfo;
+import org.junit.Before;
import org.junit.Test;
import org.apache.hadoop.ipc.TestSaslRPC.TestTokenSecretManager;
import org.apache.hadoop.ipc.TestSaslRPC.TestTokenIdentifier;
@@ -58,7 +59,7 @@ public class TestDoAsEffectiveUser {
GROUP2_NAME };
private static final String ADDRESS = "0.0.0.0";
private TestProtocol proxy;
- private static Configuration masterConf = new Configuration();
+ private static final Configuration masterConf = new Configuration();
public static final Log LOG = LogFactory
@@ -70,6 +71,10 @@ public class TestDoAsEffectiveUser {
"RULE:[2:$1@$0](.*@HADOOP.APACHE.ORG)s/@.*//" +
"RULE:[1:$1@$0](.*@HADOOP.APACHE.ORG)s/@.*//"
+ "DEFAULT");
+ }
+
+ @Before
+ public void setMasterConf() {
UserGroupInformation.setConfiguration(masterConf);
}
Propchange: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/
------------------------------------------------------------------------------
--- svn:ignore (original)
+++ svn:ignore Thu Sep 19 23:42:10 2013
@@ -1 +1,4 @@
+.classpath
+.project
+.settings
target
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/pom.xml
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/pom.xml?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/pom.xml (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/pom.xml Thu Sep 19 23:42:10 2013
@@ -38,7 +38,7 @@
<dependency>
<groupId>org.apache.directory.server</groupId>
<artifactId>apacheds-all</artifactId>
- <version>2.0.0-M14</version>
+ <version>2.0.0-M15</version>
<scope>compile</scope>
</dependency>
<dependency>
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/main/java/org/apache/hadoop/minikdc/MiniKdc.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/main/java/org/apache/hadoop/minikdc/MiniKdc.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/main/java/org/apache/hadoop/minikdc/MiniKdc.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/main/java/org/apache/hadoop/minikdc/MiniKdc.java Thu Sep 19 23:42:10 2013
@@ -37,7 +37,7 @@ import org.apache.directory.server.core.
import org.apache.directory.server.core.partition.ldif.LdifPartition;
import org.apache.directory.server.kerberos.kdc.KdcServer;
import org.apache.directory.server.kerberos.shared.crypto.encryption.KerberosKeyFactory;
-import org.apache.directory.server.kerberos.shared.keytab.HackedKeytab;
+import org.apache.directory.server.kerberos.shared.keytab.Keytab;
import org.apache.directory.server.kerberos.shared.keytab.KeytabEntry;
import org.apache.directory.server.protocol.shared.transport.TcpTransport;
import org.apache.directory.server.protocol.shared.transport.UdpTransport;
@@ -59,6 +59,7 @@ import java.io.FileReader;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.io.StringReader;
+import java.lang.reflect.Method;
import java.net.InetAddress;
import java.net.ServerSocket;
import java.text.MessageFormat;
@@ -432,6 +433,17 @@ public class MiniKdc {
System.setProperty("sun.security.krb5.debug", conf.getProperty(DEBUG,
"false"));
+
+ // refresh the config
+ Class<?> classRef;
+ if (System.getProperty("java.vendor").contains("IBM")) {
+ classRef = Class.forName("com.ibm.security.krb5.internal.Config");
+ } else {
+ classRef = Class.forName("sun.security.krb5.Config");
+ }
+ Method refreshMethod = classRef.getMethod("refresh", new Class[0]);
+ refreshMethod.invoke(classRef, new Object[0]);
+
LOG.info("MiniKdc listening at port: {}", getPort());
LOG.info("MiniKdc setting JVM krb5.conf to: {}",
krb5conf.getAbsolutePath());
@@ -514,7 +526,7 @@ public class MiniKdc {
public void createPrincipal(File keytabFile, String ... principals)
throws Exception {
String generatedPassword = UUID.randomUUID().toString();
- HackedKeytab keytab = new HackedKeytab();
+ Keytab keytab = new Keytab();
List<KeytabEntry> entries = new ArrayList<KeytabEntry>();
for (String principal : principals) {
createPrincipal(principal, generatedPassword);
@@ -529,6 +541,6 @@ public class MiniKdc {
}
}
keytab.setEntries(entries);
- keytab.write(keytabFile, principals.length);
+ keytab.write(keytabFile);
}
}
\ No newline at end of file
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/test/java/org/apache/hadoop/minikdc/TestMiniKdc.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/test/java/org/apache/hadoop/minikdc/TestMiniKdc.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/test/java/org/apache/hadoop/minikdc/TestMiniKdc.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-minikdc/src/test/java/org/apache/hadoop/minikdc/TestMiniKdc.java Thu Sep 19 23:42:10 2013
@@ -30,7 +30,11 @@ import javax.security.auth.login.Configu
import javax.security.auth.login.LoginContext;
import java.io.File;
import java.security.Principal;
-import java.util.*;
+import java.util.Set;
+import java.util.Map;
+import java.util.HashSet;
+import java.util.HashMap;
+import java.util.Arrays;
public class TestMiniKdc extends KerberosSecurityTestcase {
@@ -137,7 +141,7 @@ public class TestMiniKdc extends Kerbero
subject.getPrincipals().iterator().next().getClass());
Assert.assertEquals(principal + "@" + kdc.getRealm(),
subject.getPrincipals().iterator().next().getName());
- loginContext.login();
+ loginContext.logout();
//server login
subject = new Subject(false, principals, new HashSet<Object>(),
@@ -151,7 +155,7 @@ public class TestMiniKdc extends Kerbero
subject.getPrincipals().iterator().next().getClass());
Assert.assertEquals(principal + "@" + kdc.getRealm(),
subject.getPrincipals().iterator().next().getName());
- loginContext.login();
+ loginContext.logout();
} finally {
if (loginContext != null) {
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/pom.xml
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/pom.xml?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/pom.xml (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/pom.xml Thu Sep 19 23:42:10 2013
@@ -95,4 +95,50 @@
<version>11.0.2</version>
</dependency>
</dependencies>
+
+
+ <profiles>
+ <profile>
+ <id>dist</id>
+ <activation>
+ <activeByDefault>false</activeByDefault>
+ </activation>
+ <build>
+ <plugins>
+ <plugin>
+ <groupId>org.apache.maven.plugins</groupId>
+ <artifactId>maven-assembly-plugin</artifactId>
+ <dependencies>
+ <dependency>
+ <groupId>org.apache.hadoop</groupId>
+ <artifactId>hadoop-assemblies</artifactId>
+ <version>${project.version}</version>
+ </dependency>
+ </dependencies>
+ <executions>
+ <execution>
+ <id>dist</id>
+ <phase>package</phase>
+ <goals>
+ <goal>single</goal>
+ </goals>
+ <configuration>
+ <finalName>${project.artifactId}-${project.version}</finalName>
+ <appendAssemblyId>false</appendAssemblyId>
+ <attach>false</attach>
+ <!--<descriptorRefs>
+ <descriptorRef>hadoop-nfs-dist</descriptorRef>
+ </descriptorRefs>-->
+ <descriptors>
+ <descriptor>../../hadoop-assemblies/src/main/resources/assemblies/hadoop-nfs-dist.xml</descriptor>
+ </descriptors>
+ </configuration>
+ </execution>
+ </executions>
+ </plugin>
+ </plugins>
+ </build>
+ </profile>
+ </profiles>
+
</project>
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/mount/MountResponse.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/mount/MountResponse.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/mount/MountResponse.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/mount/MountResponse.java Thu Sep 19 23:42:10 2013
@@ -19,9 +19,10 @@ package org.apache.hadoop.mount;
import java.util.List;
+import org.apache.hadoop.nfs.NfsExports;
import org.apache.hadoop.oncrpc.RpcAcceptedReply;
import org.apache.hadoop.oncrpc.XDR;
-import org.apache.hadoop.oncrpc.RpcAuthInfo.AuthFlavor;
+import org.apache.hadoop.oncrpc.security.RpcAuthInfo.AuthFlavor;
/**
* Helper class for sending MountResponse
@@ -59,15 +60,28 @@ public class MountResponse {
xdr.writeBoolean(false); // Value follows no
return xdr;
}
-
+
/** Response for RPC call {@link MountInterface.MNTPROC#EXPORT} */
- public static XDR writeExportList(XDR xdr, int xid, List<String> exports) {
+ public static XDR writeExportList(XDR xdr, int xid, List<String> exports,
+ List<NfsExports> hostMatcher) {
+ assert (exports.size() == hostMatcher.size());
+
RpcAcceptedReply.voidReply(xdr, xid);
- for (String export : exports) {
+ for (int i = 0; i < exports.size(); i++) {
xdr.writeBoolean(true); // Value follows - yes
- xdr.writeString(export);
- xdr.writeInt(0);
+ xdr.writeString(exports.get(i));
+
+ // List host groups
+ String[] hostGroups = hostMatcher.get(i).getHostGroupList();
+ if (hostGroups.length > 0) {
+ for (int j = 0; j < hostGroups.length; j++) {
+ xdr.writeBoolean(true); // Value follows - yes
+ xdr.writeVariableOpaque(hostGroups[j].getBytes());
+ }
+ }
+ xdr.writeBoolean(false); // Value follows - no more group
}
+
xdr.writeBoolean(false); // Value follows - no
return xdr;
}
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/IdUserGroup.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/IdUserGroup.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/IdUserGroup.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/IdUserGroup.java Thu Sep 19 23:42:10 2013
@@ -147,7 +147,7 @@ public class IdUserGroup {
synchronized public String getUserName(int uid, String unknown) {
checkAndUpdateMaps();
- String uname = uidNameMap.get(Integer.valueOf(uid));
+ String uname = uidNameMap.get(uid);
if (uname == null) {
uname = unknown;
}
@@ -156,7 +156,7 @@ public class IdUserGroup {
synchronized public String getGroupName(int gid, String unknown) {
checkAndUpdateMaps();
- String gname = gidNameMap.get(Integer.valueOf(gid));
+ String gname = gidNameMap.get(gid);
if (gname == null) {
gname = unknown;
}
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Constant.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Constant.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Constant.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Constant.java Thu Sep 19 23:42:10 2013
@@ -189,4 +189,23 @@ public class Nfs3Constant {
public final static int CREATE_UNCHECKED = 0;
public final static int CREATE_GUARDED = 1;
public final static int CREATE_EXCLUSIVE = 2;
+
+ public static final String EXPORTS_ALLOWED_HOSTS_SEPARATOR = ";";
+ /** Allowed hosts for nfs exports */
+ public static final String EXPORTS_ALLOWED_HOSTS_KEY = "dfs.nfs.exports.allowed.hosts";
+ public static final String EXPORTS_ALLOWED_HOSTS_KEY_DEFAULT = "* rw";
+ /** Size for nfs exports cache */
+ public static final String EXPORTS_CACHE_SIZE_KEY = "dfs.nfs.exports.cache.size";
+ public static final int EXPORTS_CACHE_SIZE_DEFAULT = 512;
+ /** Expiration time for nfs exports cache entry */
+ public static final String EXPORTS_CACHE_EXPIRYTIME_MILLIS_KEY = "dfs.nfs.exports.cache.expirytime.millis";
+ public static final long EXPORTS_CACHE_EXPIRYTIME_MILLIS_DEFAULT = 15 * 60 * 1000; // 15 min
+
+ public static final String FILE_DUMP_DIR_KEY = "dfs.nfs3.dump.dir";
+ public static final String FILE_DUMP_DIR_DEFAULT = "/tmp/.hdfs-nfs";
+ public static final String ENABLE_FILE_DUMP_KEY = "dfs.nfs3.enableDump";
+ public static final boolean ENABLE_FILE_DUMP_DEFAULT = true;
+
+ public final static String UNKNOWN_USER = "nobody";
+ public final static String UNKNOWN_GROUP = "nobody";
}
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Interface.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Interface.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Interface.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Interface.java Thu Sep 19 23:42:10 2013
@@ -17,67 +17,86 @@
*/
package org.apache.hadoop.nfs.nfs3;
+import java.net.InetAddress;
+
import org.apache.hadoop.nfs.nfs3.response.NFS3Response;
-import org.apache.hadoop.oncrpc.RpcAuthSys;
import org.apache.hadoop.oncrpc.XDR;
+import org.apache.hadoop.oncrpc.security.SecurityHandler;
import org.jboss.netty.channel.Channel;
/**
* RPC procedures as defined in RFC 1813.
*/
public interface Nfs3Interface {
-
+
/** NULL: Do nothing */
public NFS3Response nullProcedure();
-
+
/** GETATTR: Get file attributes */
- public NFS3Response getattr(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response getattr(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** SETATTR: Set file attributes */
- public NFS3Response setattr(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response setattr(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** LOOKUP: Lookup filename */
- public NFS3Response lookup(XDR xdr, RpcAuthSys authSys);
-
- /** ACCESS: Check access permission */
- public NFS3Response access(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response lookup(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** ACCESS: Check access permission */
+ public NFS3Response access(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** READ: Read from file */
- public NFS3Response read(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response read(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** WRITE: Write to file */
- public NFS3Response write(XDR xdr, Channel channel, int xid, RpcAuthSys authSys);
-
- /** CREATE: Create a file */
- public NFS3Response create(XDR xdr, RpcAuthSys authSys);
-
- /** MKDIR: Create a directory */
- public NFS3Response mkdir(XDR xdr, RpcAuthSys authSys);
-
- /** REMOVE: Remove a file */
- public NFS3Response remove(XDR xdr, RpcAuthSys authSys);
-
- /** RMDIR: Remove a directory */
- public NFS3Response rmdir(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response write(XDR xdr, Channel channel, int xid,
+ SecurityHandler securityHandler, InetAddress client);
+
+ /** CREATE: Create a file */
+ public NFS3Response create(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** MKDIR: Create a directory */
+ public NFS3Response mkdir(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** REMOVE: Remove a file */
+ public NFS3Response remove(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** RMDIR: Remove a directory */
+ public NFS3Response rmdir(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** RENAME: Rename a file or directory */
- public NFS3Response rename(XDR xdr, RpcAuthSys authSys);
-
- /** SYMLINK: Create a symbolic link */
- public NFS3Response symlink(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response rename(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** SYMLINK: Create a symbolic link */
+ public NFS3Response symlink(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** READDIR: Read From directory */
- public NFS3Response readdir(XDR xdr, RpcAuthSys authSys);
-
- /** FSSTAT: Get dynamic file system information */
- public NFS3Response fsstat(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response readdir(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** FSSTAT: Get dynamic file system information */
+ public NFS3Response fsstat(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** FSINFO: Get static file system information */
- public NFS3Response fsinfo(XDR xdr, RpcAuthSys authSys);
-
+ public NFS3Response fsinfo(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
/** PATHCONF: Retrieve POSIX information */
- public NFS3Response pathconf(XDR xdr, RpcAuthSys authSys);
-
- /** COMMIT: Commit cached data on a server to stable storage */
- public NFS3Response commit(XDR xdr, RpcAuthSys authSys);
+ public NFS3Response pathconf(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
+
+ /** COMMIT: Commit cached data on a server to stable storage */
+ public NFS3Response commit(XDR xdr, SecurityHandler securityHandler,
+ InetAddress client);
}
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIR3Response.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIR3Response.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIR3Response.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIR3Response.java Thu Sep 19 23:42:10 2013
@@ -17,12 +17,14 @@
*/
package org.apache.hadoop.nfs.nfs3.response;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.List;
+
import org.apache.hadoop.nfs.nfs3.Nfs3FileAttributes;
import org.apache.hadoop.nfs.nfs3.Nfs3Status;
import org.apache.hadoop.oncrpc.XDR;
-import com.google.common.collect.ObjectArrays;
-
/**
* READDIR3 Response
*/
@@ -56,12 +58,11 @@ public class READDIR3Response extends NF
}
public static class DirList3 {
- final Entry3 entries[];
+ final List<Entry3> entries;
final boolean eof;
public DirList3(Entry3[] entries, boolean eof) {
- this.entries = ObjectArrays.newArray(entries, entries.length);
- System.arraycopy(this.entries, 0, entries, 0, entries.length);
+ this.entries = Collections.unmodifiableList(Arrays.asList(entries));
this.eof = eof;
}
}
@@ -102,12 +103,11 @@ public class READDIR3Response extends NF
if (getStatus() == Nfs3Status.NFS3_OK) {
xdr.writeLongAsHyper(cookieVerf);
- Entry3[] f = dirList.entries;
- for (int i = 0; i < f.length; i++) {
+ for (Entry3 e : dirList.entries) {
xdr.writeBoolean(true); // Value follows
- xdr.writeLongAsHyper(f[i].getFileId());
- xdr.writeString(f[i].getName());
- xdr.writeLongAsHyper(f[i].getCookie());
+ xdr.writeLongAsHyper(e.getFileId());
+ xdr.writeString(e.getName());
+ xdr.writeLongAsHyper(e.getCookie());
}
xdr.writeBoolean(false);
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIRPLUS3Response.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIRPLUS3Response.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIRPLUS3Response.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/response/READDIRPLUS3Response.java Thu Sep 19 23:42:10 2013
@@ -17,13 +17,15 @@
*/
package org.apache.hadoop.nfs.nfs3.response;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.List;
+
import org.apache.hadoop.nfs.nfs3.FileHandle;
import org.apache.hadoop.nfs.nfs3.Nfs3FileAttributes;
import org.apache.hadoop.nfs.nfs3.Nfs3Status;
import org.apache.hadoop.oncrpc.XDR;
-import com.google.common.collect.ObjectArrays;
-
/**
* READDIRPLUS3 Response
*/
@@ -60,16 +62,15 @@ public class READDIRPLUS3Response exten
}
public static class DirListPlus3 {
- EntryPlus3 entries[];
+ List<EntryPlus3> entries;
boolean eof;
public DirListPlus3(EntryPlus3[] entries, boolean eof) {
- this.entries = ObjectArrays.newArray(entries, entries.length);
- System.arraycopy(this.entries, 0, entries, 0, entries.length);
+ this.entries = Collections.unmodifiableList(Arrays.asList(entries));
this.eof = eof;
}
- EntryPlus3[] getEntries() {
+ List<EntryPlus3> getEntries() {
return entries;
}
@@ -101,10 +102,9 @@ public class READDIRPLUS3Response exten
if (getStatus() == Nfs3Status.NFS3_OK) {
out.writeLongAsHyper(cookieVerf);
- EntryPlus3[] f = dirListPlus.getEntries();
- for (int i = 0; i < f.length; i++) {
+ for (EntryPlus3 f : dirListPlus.getEntries()) {
out.writeBoolean(true); // next
- f[i].seralize(out);
+ f.seralize(out);
}
out.writeBoolean(false);
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcAcceptedReply.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcAcceptedReply.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcAcceptedReply.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcAcceptedReply.java Thu Sep 19 23:42:10 2013
@@ -17,7 +17,9 @@
*/
package org.apache.hadoop.oncrpc;
-import org.apache.hadoop.oncrpc.RpcAuthInfo.AuthFlavor;
+import org.apache.hadoop.oncrpc.security.Verifier;
+import org.apache.hadoop.oncrpc.security.RpcAuthInfo;
+import org.apache.hadoop.oncrpc.security.RpcAuthInfo.AuthFlavor;
/**
* Represents RPC message MSG_ACCEPTED reply body. See RFC 1831 for details.
@@ -54,7 +56,7 @@ public class RpcAcceptedReply extends Rp
public static RpcAcceptedReply read(int xid, RpcMessage.Type messageType,
ReplyState replyState, XDR xdr) {
- RpcAuthInfo verifier = RpcAuthInfo.read(xdr);
+ Verifier verifier = Verifier.readFlavorAndVerifier(xdr);
AcceptState acceptState = AcceptState.fromValue(xdr.readInt());
return new RpcAcceptedReply(xid, messageType, replyState, verifier,
acceptState);
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcCall.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcCall.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcCall.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcCall.java Thu Sep 19 23:42:10 2013
@@ -19,6 +19,8 @@ package org.apache.hadoop.oncrpc;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
+import org.apache.hadoop.oncrpc.security.Credentials;
+import org.apache.hadoop.oncrpc.security.Verifier;
/**
* Represents an RPC message of type RPC call as defined in RFC 1831
@@ -30,11 +32,12 @@ public class RpcCall extends RpcMessage
private final int program;
private final int version;
private final int procedure;
- private final RpcAuthInfo credential;
- private final RpcAuthInfo verifier;
+ private final Credentials credential;
+ private final Verifier verifier;
- protected RpcCall(int xid, RpcMessage.Type messageType, int rpcVersion, int program,
- int version, int procedure, RpcAuthInfo credential, RpcAuthInfo verifier) {
+ protected RpcCall(int xid, RpcMessage.Type messageType, int rpcVersion,
+ int program, int version, int procedure, Credentials credential,
+ Verifier verifier) {
super(xid, messageType);
this.rpcVersion = rpcVersion;
this.program = program;
@@ -79,19 +82,19 @@ public class RpcCall extends RpcMessage
return procedure;
}
- public RpcAuthInfo getCredential() {
+ public Credentials getCredential() {
return credential;
}
- public RpcAuthInfo getVerifier() {
+ public Verifier getVerifier() {
return verifier;
}
public static RpcCall read(XDR xdr) {
return new RpcCall(xdr.readInt(), RpcMessage.Type.fromValue(xdr.readInt()),
- xdr.readInt(), xdr.readInt(),
- xdr.readInt(), xdr.readInt(), RpcAuthInfo.read(xdr),
- RpcAuthInfo.read(xdr));
+ xdr.readInt(), xdr.readInt(), xdr.readInt(), xdr.readInt(),
+ Credentials.readFlavorAndCredentials(xdr),
+ Verifier.readFlavorAndVerifier(xdr));
}
public static void write(XDR out, int xid, int program, int progVersion,
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcDeniedReply.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcDeniedReply.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcDeniedReply.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcDeniedReply.java Thu Sep 19 23:42:10 2013
@@ -17,7 +17,7 @@
*/
package org.apache.hadoop.oncrpc;
-import org.apache.hadoop.oncrpc.RpcAuthInfo.AuthFlavor;
+import org.apache.hadoop.oncrpc.security.RpcAuthInfo.AuthFlavor;
/**
* Represents RPC message MSG_DENIED reply body. See RFC 1831 for details.
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/XDR.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/XDR.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/XDR.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/XDR.java Thu Sep 19 23:42:10 2013
@@ -280,7 +280,7 @@ public class XDR {
public byte[] readVariableOpaque() {
int size = this.readInt();
- return size != 0 ? this.readFixedOpaque(size) : null;
+ return size != 0 ? this.readFixedOpaque(size) : new byte[0];
}
public void skipVariableOpaque() {
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/portmap/PortmapRequest.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/portmap/PortmapRequest.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/portmap/PortmapRequest.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/portmap/PortmapRequest.java Thu Sep 19 23:42:10 2013
@@ -17,10 +17,14 @@
*/
package org.apache.hadoop.portmap;
-import org.apache.hadoop.oncrpc.RpcAuthInfo.AuthFlavor;
import org.apache.hadoop.oncrpc.RpcCall;
import org.apache.hadoop.oncrpc.RpcUtil;
import org.apache.hadoop.oncrpc.XDR;
+import org.apache.hadoop.oncrpc.security.CredentialsNone;
+import org.apache.hadoop.oncrpc.security.Credentials;
+import org.apache.hadoop.oncrpc.security.Verifier;
+import org.apache.hadoop.oncrpc.security.VerifierNone;
+import org.apache.hadoop.oncrpc.security.RpcAuthInfo.AuthFlavor;
import org.apache.hadoop.portmap.PortmapInterface.Procedure;
/**
@@ -38,9 +42,11 @@ public class PortmapRequest {
RpcProgramPortmap.PROGRAM, RpcProgramPortmap.VERSION,
Procedure.PMAPPROC_SET.getValue());
request.writeInt(AuthFlavor.AUTH_NONE.getValue());
- request.writeInt(0);
- request.writeInt(0);
- request.writeInt(0);
+ Credentials credential = new CredentialsNone();
+ credential.write(request);
+ request.writeInt(AuthFlavor.AUTH_NONE.getValue());
+ Verifier verifier = new VerifierNone();
+ verifier.write(request);
return mapping.serialize(request);
}
}
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcAcceptedReply.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcAcceptedReply.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcAcceptedReply.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcAcceptedReply.java Thu Sep 19 23:42:10 2013
@@ -20,8 +20,9 @@ package org.apache.hadoop.oncrpc;
import static org.junit.Assert.assertEquals;
import org.apache.hadoop.oncrpc.RpcAcceptedReply.AcceptState;
-import org.apache.hadoop.oncrpc.RpcAuthInfo.AuthFlavor;
import org.apache.hadoop.oncrpc.RpcReply.ReplyState;
+import org.apache.hadoop.oncrpc.security.Verifier;
+import org.apache.hadoop.oncrpc.security.VerifierNone;
import org.junit.Test;
/**
@@ -45,7 +46,7 @@ public class TestRpcAcceptedReply {
@Test
public void testConstructor() {
- RpcAuthInfo verifier = new RpcAuthInfo(AuthFlavor.AUTH_NONE, new byte[0]);
+ Verifier verifier = new VerifierNone();
RpcAcceptedReply reply = new RpcAcceptedReply(0, RpcMessage.Type.RPC_REPLY,
ReplyState.MSG_ACCEPTED, verifier, AcceptState.SUCCESS);
assertEquals(0, reply.getXid());
Modified: hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcCall.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcCall.java?rev=1524865&r1=1524864&r2=1524865&view=diff
==============================================================================
--- hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcCall.java (original)
+++ hadoop/common/branches/HDFS-4949/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestRpcCall.java Thu Sep 19 23:42:10 2013
@@ -17,8 +17,12 @@
*/
package org.apache.hadoop.oncrpc;
-import org.apache.hadoop.oncrpc.RpcAuthInfo.AuthFlavor;
import static org.junit.Assert.assertEquals;
+
+import org.apache.hadoop.oncrpc.security.CredentialsNone;
+import org.apache.hadoop.oncrpc.security.Credentials;
+import org.apache.hadoop.oncrpc.security.Verifier;
+import org.apache.hadoop.oncrpc.security.VerifierNone;
import org.junit.Test;
/**
@@ -28,8 +32,8 @@ public class TestRpcCall {
@Test
public void testConstructor() {
- RpcAuthInfo credential = new RpcAuthInfo(AuthFlavor.AUTH_NONE, new byte[0]);
- RpcAuthInfo verifier = new RpcAuthInfo(AuthFlavor.AUTH_NONE, new byte[0]);
+ Credentials credential = new CredentialsNone();
+ Verifier verifier = new VerifierNone();
int rpcVersion = RpcCall.RPC_VERSION;
int program = 2;
int version = 3;