You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-issues@hadoop.apache.org by "Kihwal Lee (Jira)" <ji...@apache.org> on 2021/02/17 19:05:00 UTC
[jira] [Resolved] (HADOOP-17533) Server IPC version 9 cannot
communicate with client version 4
[ https://issues.apache.org/jira/browse/HADOOP-17533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kihwal Lee resolved HADOOP-17533.
---------------------------------
Resolution: Invalid
> Server IPC version 9 cannot communicate with client version 4
> -------------------------------------------------------------
>
> Key: HADOOP-17533
> URL: https://issues.apache.org/jira/browse/HADOOP-17533
> Project: Hadoop Common
> Issue Type: Bug
> Reporter: Mariia
> Priority: Major
> Labels: hadoop, hdfs, java, maven
>
> `I want to connect to hdfs using java jast like this
> _String url = "hdfs://c7301.ambari.apache.org:8020/file.txt";_
> _FileSystem fs = null;_
> _InputStream in = null;_
> _try {_
> _Configuration conf = new Configuration();_
> _fs = FileSystem.get(URI.create(url), conf, "admin");_
> _in = fs.open(new Path(url));_
> _IOUtils.copyBytes(in, System.out, 4096, false);_
> _} catch (Exception e) {_
> _e.printStackTrace();_
> _} finally {_
> _IOUtils.closeStream(fs);_
> _}_
> *Error that i got*
> [2021-02-17 20:02:06,115] ERROR PriviledgedActionException as:admin cause:org.apache.hadoop.ipc.RemoteException: *Server IPC version 9 cannot communicate with* *client version 4 (*org.apache.hadoop.security.UserGroupInformation:1124)
> org.apache.hadoop.ipc.RemoteException: *Server IPC version 9 cannot communicate with client version 4*
> at org.apache.hadoop.ipc.Client.call(Client.java:1070)
> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
> at com.sun.proxy.$Proxy4.getProtocolVersion(Unknown Source)
> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
> at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119)
> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238)
> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203)
> at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386)
> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)
> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
> at org.apache.hadoop.fs.FileSystem$1.run(FileSystem.java:117)
> at org.apache.hadoop.fs.FileSystem$1.run(FileSystem.java:115)
> at java.base/java.security.AccessController.doPrivileged(Native Method)
> at java.base/javax.security.auth.Subject.doAs(Subject.java:423)
> at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:115)
> at Main.main(Main.java:38)
> \{{ I tried different solutions to the problem, but nothing helped. }}
> *Its my pom.xml file*
>
> <?xml version="1.0" encoding="UTF-8"?>
> <project xmlns="http://maven.apache.org/POM/4.0.0"
> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 [http://maven.apache.org/xsd/maven-4.0.0.xsd]">
> <modelVersion>4.0.0</modelVersion>
> <groupId>org.example</groupId>
> <artifactId>producer</artifactId>
> <version>1.0-SNAPSHOT</version>
> <properties>
> <maven.compiler.source>11</maven.compiler.source>
> <maven.compiler.target>11</maven.compiler.target>
> </properties>
> <dependencies>
> <dependency>
> <groupId>org.apache.kafka</groupId>
> <artifactId>kafka-clients</artifactId>
> <version>0.10.0.0</version>
> </dependency>
> <dependency>
> <groupId>org.apache.hadoop</groupId>
> <artifactId>hadoop-common</artifactId>
> <version>3.2.0</version>
> </dependency>
> <dependency>
> <groupId>org.apache.hadoop</groupId>
> <artifactId>hadoop-hdfs</artifactId>
> <version>3.2.0</version>
> </dependency>
> <dependency>
> <groupId>org.apache.hadoop</groupId>
> <artifactId>hadoop-yarn-common</artifactId>
> <version>3.2.0</version>
> </dependency>
> <dependency>
> <groupId>org.apache.hadoop</groupId>
> <artifactId>hadoop-mapreduce-client-common</artifactId>
> <version>3.2.0</version>
> </dependency>
> <dependency>
> <groupId>org.apache.hadoop</groupId>
> <artifactId>hadoop-mapreduce-client-core</artifactId>
> <version>3.2.0</version>
> </dependency>
> </dependencies>
> <build>
> <plugins>
> <plugin>
> <groupId>org.apache.maven.plugins</groupId>
> <artifactId>maven-shade-plugin</artifactId>
> <version>3.2.4</version>
> <executions>
> <execution>
> <phase>package</phase>
> <goals>
> <goal>shade</goal>
> </goals>
> <configuration>
> <transformers>
> <transformer
> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
> <mainClass>Main</mainClass>
> </transformer>
> </transformers>
> </configuration>
> </execution>
> </executions>
> </plugin>
> </plugins>
> </build>
> </project>
> \{{}}
> *And its version of hdfs*
> Hadoop 3.1.1.3.1.4.0-315 Source code repository git@github.com:hortonworks/hadoop.git -r 58d0fd3d8ce58b10149da3c717c45e5e57a60d14 Compiled by jenkins on 2019-08-23T05:15Z Compiled with protoc 2.5.0 From source with checksum fcbd146ffa6d48fef0ed81332f9d6f0 This command was run using /usr/ddp/3.1.4.0-315/hadoop/hadoop-common-3.1.1.3.1.4.0-315.jar
>
> if someone knew a similar problem, please help
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-issues-help@hadoop.apache.org