You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@camel.apache.org by ac...@apache.org on 2019/08/05 09:48:26 UTC

[camel] 02/07: CAMEL-13792 - Rename components to default names, Camel-hdfs2 to camel-hdfs - updated references

This is an automated email from the ASF dual-hosted git repository.

acosentino pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/camel.git

commit e0ac1da5d30c0de57549a9cb07a08a0d54b93414
Author: Andrea Cosentino <an...@gmail.com>
AuthorDate: Mon Aug 5 10:38:30 2019 +0200

    CAMEL-13792 - Rename components to default names, Camel-hdfs2 to camel-hdfs - updated references
---
 apache-camel/pom.xml                               |  4 +--
 apache-camel/src/main/descriptors/common-bin.xml   |  4 +--
 bom/camel-bom/pom.xml                              |  4 +--
 .../{hdfs2-component.adoc => hdfs-component.adoc}  | 20 ++++++------
 .../camel/component/hdfs2/HdfsComponent.java       |  2 +-
 .../camel/component/hdfs2/HdfsConfiguration.java   |  2 +-
 .../apache/camel/component/hdfs2/HdfsEndpoint.java |  2 +-
 .../camel/component/hdfs2/FromFileToHdfsTest.java  |  2 +-
 .../camel/component/hdfs2/HdfsConsumerTest.java    | 36 +++++++++++-----------
 .../component/hdfs2/HdfsProducerConsumerTest.java  |  4 +--
 .../component/hdfs2/HdfsProducerSplitTest.java     | 10 +++---
 .../camel/component/hdfs2/HdfsProducerTest.java    | 26 ++++++++--------
 .../hdfs2/integration/HdfsAppendTest.java          |  8 ++---
 .../HdfsProducerConsumerIntegrationTest.java       | 12 ++++----
 .../src/test/resources/log4j2.properties           |  2 +-
 components/pom.xml                                 |  2 +-
 .../endpoint/dsl/HdfsEndpointBuilderFactory.java   | 14 ++++-----
 parent/pom.xml                                     |  4 +--
 .../springboot/HdfsComponentAutoConfiguration.java |  2 +-
 platforms/spring-boot/components-starter/pom.xml   |  1 -
 20 files changed, 80 insertions(+), 81 deletions(-)

diff --git a/apache-camel/pom.xml b/apache-camel/pom.xml
index ba58248..df4ab2b 100644
--- a/apache-camel/pom.xml
+++ b/apache-camel/pom.xml
@@ -679,7 +679,7 @@
     </dependency>
     <dependency>
       <groupId>org.apache.camel</groupId>
-      <artifactId>camel-hdfs2</artifactId>
+      <artifactId>camel-hdfs</artifactId>
       <version>${project.version}</version>
     </dependency>
     <dependency>
@@ -2188,7 +2188,7 @@
     </dependency>
     <dependency>
       <groupId>org.apache.camel</groupId>
-      <artifactId>camel-hdfs2-starter</artifactId>
+      <artifactId>camel-hdfs-starter</artifactId>
       <version>${project.version}</version>
     </dependency>
     <dependency>
diff --git a/apache-camel/src/main/descriptors/common-bin.xml b/apache-camel/src/main/descriptors/common-bin.xml
index 90abf7a..131e951 100644
--- a/apache-camel/src/main/descriptors/common-bin.xml
+++ b/apache-camel/src/main/descriptors/common-bin.xml
@@ -156,7 +156,7 @@
         <include>org.apache.camel:camel-guava-eventbus</include>
         <include>org.apache.camel:camel-hazelcast</include>
         <include>org.apache.camel:camel-hbase</include>
-        <include>org.apache.camel:camel-hdfs2</include>
+        <include>org.apache.camel:camel-hdfs</include>
         <include>org.apache.camel:camel-hipchat</include>
         <include>org.apache.camel:camel-hl7</include>
         <include>org.apache.camel:camel-http</include>
@@ -497,7 +497,7 @@
         <include>org.apache.camel:camel-guava-eventbus-starter</include>
         <include>org.apache.camel:camel-hazelcast-starter</include>
         <include>org.apache.camel:camel-hbase-starter</include>
-        <include>org.apache.camel:camel-hdfs2-starter</include>
+        <include>org.apache.camel:camel-hdfs-starter</include>
         <include>org.apache.camel:camel-hipchat-starter</include>
         <include>org.apache.camel:camel-hl7-starter</include>
         <include>org.apache.camel:camel-http-common-starter</include>
diff --git a/bom/camel-bom/pom.xml b/bom/camel-bom/pom.xml
index f1ed1c8..7033723 100644
--- a/bom/camel-bom/pom.xml
+++ b/bom/camel-bom/pom.xml
@@ -1244,12 +1244,12 @@
       </dependency>
       <dependency>
         <groupId>org.apache.camel</groupId>
-        <artifactId>camel-hdfs2</artifactId>
+        <artifactId>camel-hdfs</artifactId>
         <version>${project.version}</version>
       </dependency>
       <dependency>
         <groupId>org.apache.camel</groupId>
-        <artifactId>camel-hdfs2-starter</artifactId>
+        <artifactId>camel-hdfs-starter</artifactId>
         <version>${project.version}</version>
       </dependency>
       <dependency>
diff --git a/components/camel-hdfs/src/main/docs/hdfs2-component.adoc b/components/camel-hdfs/src/main/docs/hdfs-component.adoc
similarity index 97%
rename from components/camel-hdfs/src/main/docs/hdfs2-component.adoc
rename to components/camel-hdfs/src/main/docs/hdfs-component.adoc
index f42ba0e..b8880ef 100644
--- a/components/camel-hdfs/src/main/docs/hdfs2-component.adoc
+++ b/components/camel-hdfs/src/main/docs/hdfs-component.adoc
@@ -1,9 +1,9 @@
-[[hdfs2-component]]
-= HDFS2 Component
+[[hdfs-component]]
+= HDFS Component
 
 *Available as of Camel version 2.14*
 
-The HDFS2 component enables you to read and write messages from/to an
+The HDFS component enables you to read and write messages from/to an
 HDFS file system using Hadoop 2.x. HDFS is the distributed file system
 at the heart of http://hadoop.apache.org[Hadoop].
 
@@ -24,7 +24,7 @@ for this component:
 
 [source,java]
 ----------------------------------------
-hdfs2://hostname[:port][/path][?options]
+hdfs://hostname[:port][/path][?options]
 ----------------------------------------
 
 You can append query options to the URI in the following format,
@@ -41,7 +41,7 @@ a different file per split named using the configured
 UuidGenerator.
 
 
-When consuming from hdfs2 then in normal mode, a file is split into
+When consuming from hdfs then in normal mode, a file is split into
 chunks, producing a message per chunk. You can configure the size of the
 chunk using the chunkSize option. If you want to read from hdfs and
 write to a regular file using the file component, then you can use the
@@ -54,7 +54,7 @@ fileMode=Append to append each of the chunks together.
 
 
 // component options: START
-The HDFS2 component supports 3 options, which are listed below.
+The HDFS component supports 3 options, which are listed below.
 
 
 
@@ -73,10 +73,10 @@ The HDFS2 component supports 3 options, which are listed below.
 
 
 // endpoint options: START
-The HDFS2 endpoint is configured using URI syntax:
+The HDFS endpoint is configured using URI syntax:
 
 ----
-hdfs2:hostName:port/path
+hdfs:hostName:port/path
 ----
 
 with the following path and query parameters:
@@ -223,7 +223,7 @@ for example:
 
 [source,java]
 -----------------------------------------------------------------
-hdfs2://localhost/tmp/simple-file?splitStrategy=IDLE:1000,BYTES:5
+hdfs://localhost/tmp/simple-file?splitStrategy=IDLE:1000,BYTES:5
 -----------------------------------------------------------------
 
 it means: a new file is created either when it has been idle for more
@@ -249,7 +249,7 @@ split strategy.
 
 == Controlling to close file stream
 
-When using the xref:hdfs2-component.adoc[HDFS2] producer *without* a split
+When using the xref:hdfs-component.adoc[HDFS] producer *without* a split
 strategy, then the file output stream is by default closed after the
 write. However you may want to keep the stream open, and only explicitly
 close the stream later. For that you can use the header
diff --git a/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsComponent.java b/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsComponent.java
index 3ac82f2..84af870 100644
--- a/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsComponent.java
+++ b/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsComponent.java
@@ -28,7 +28,7 @@ import org.apache.hadoop.fs.FsUrlStreamHandlerFactory;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-@Component("hdfs,hdfs2")
+@Component("hdfs")
 public class HdfsComponent extends DefaultComponent {
 
     private static final Logger LOG = LoggerFactory.getLogger(HdfsComponent.class);
diff --git a/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsConfiguration.java b/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsConfiguration.java
index 174c455..24a3818 100644
--- a/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsConfiguration.java
+++ b/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsConfiguration.java
@@ -208,7 +208,7 @@ public class HdfsConfiguration {
 
     public void parseURI(URI uri) throws URISyntaxException {
         String protocol = uri.getScheme();
-        if (!protocol.equalsIgnoreCase("hdfs2")) {
+        if (!protocol.equalsIgnoreCase("hdfs")) {
             throw new IllegalArgumentException("Unrecognized protocol: " + protocol + " for uri: " + uri);
         }
         hostName = uri.getHost();
diff --git a/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsEndpoint.java b/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsEndpoint.java
index f848003d..3249c89 100644
--- a/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsEndpoint.java
+++ b/components/camel-hdfs/src/main/java/org/apache/camel/component/hdfs2/HdfsEndpoint.java
@@ -29,7 +29,7 @@ import org.apache.camel.support.ScheduledPollEndpoint;
 /**
  * For reading/writing from/to an HDFS filesystem using Hadoop 2.x.
  */
-@UriEndpoint(firstVersion = "2.14.0", scheme = "hdfs2,hdfs", title = "HDFS2", syntax = "hdfs2:hostName:port/path", label = "hadoop,file")
+@UriEndpoint(firstVersion = "2.14.0", scheme = "hdfs", title = "HDFS", syntax = "hdfs:hostName:port/path", label = "hadoop,file")
 public class HdfsEndpoint extends ScheduledPollEndpoint {
 
     @UriParam
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/FromFileToHdfsTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/FromFileToHdfsTest.java
index 79a21ae..ec5c5e0 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/FromFileToHdfsTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/FromFileToHdfsTest.java
@@ -106,7 +106,7 @@ public class FromFileToHdfsTest extends HdfsTestSupport {
             @Override
             public void configure() throws Exception {
                 from("file:target/inbox?delete=true")
-                    .to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/output.txt?fileSystemType=LOCAL");
+                    .to("hdfs:localhost/" + TEMP_DIR.toUri() + "/output.txt?fileSystemType=LOCAL");
             }
         };
     }
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsConsumerTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsConsumerTest.java
index b5bc391..99c0304 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsConsumerTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsConsumerTest.java
@@ -103,7 +103,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&chunkSize=4096&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&chunkSize=4096&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -139,10 +139,10 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=100&initialDelay=0").to("mock:result");
-                from("hdfs2://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=200&initialDelay=0").to("mock:result");
-                from("hdfs2://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=300&initialDelay=0").to("mock:result");
-                from("hdfs2://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=400&initialDelay=0").to("mock:result");
+                from("hdfs://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=100&initialDelay=0").to("mock:result");
+                from("hdfs://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=200&initialDelay=0").to("mock:result");
+                from("hdfs://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=300&initialDelay=0").to("mock:result");
+                from("hdfs://" + rootdir.toURI() + "/target/test/multiple-consumers?pattern=*.txt&fileSystemType=LOCAL&chunkSize=400&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -173,7 +173,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&chunkSize=4096&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&chunkSize=4096&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -206,7 +206,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&chunkSize=42&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&chunkSize=42&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -232,7 +232,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&chunkSize=4096&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&chunkSize=4096&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -261,7 +261,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.getParent().toUri() + "?scheduler=#myScheduler&pattern=*&fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0&readSuffix=handled")
+                from("hdfs:localhost/" + file.getParent().toUri() + "?scheduler=#myScheduler&pattern=*&fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0&readSuffix=handled")
                     .to("mock:result");
             }
         });
@@ -303,7 +303,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -336,7 +336,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -366,7 +366,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -396,7 +396,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -426,7 +426,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -456,7 +456,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -486,7 +486,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -516,7 +516,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&fileType=SEQUENCE_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
@@ -549,7 +549,7 @@ public class HdfsConsumerTest extends HdfsTestSupport {
 
         context.addRoutes(new RouteBuilder() {
             public void configure() {
-                from("hdfs2:localhost/" + file.getParent().toUri() + "?fileSystemType=LOCAL&fileType=ARRAY_FILE&initialDelay=0").to("mock:result");
+                from("hdfs:localhost/" + file.getParent().toUri() + "?fileSystemType=LOCAL&fileType=ARRAY_FILE&initialDelay=0").to("mock:result");
             }
         });
         context.start();
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerConsumerTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerConsumerTest.java
index 501ac7d..d9ed8e0 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerConsumerTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerConsumerTest.java
@@ -56,8 +56,8 @@ public class HdfsProducerConsumerTest extends HdfsTestSupport {
         context.addRoutes(new RouteBuilder() {
             @Override
             public void configure() {
-                from("direct:start").to("hdfs2:localhost/" + file.toUri() + "?fileSystemType=LOCAL&splitStrategy=BYTES:5,IDLE:1000");
-                from("hdfs2:localhost/" + file.toUri() + "?initialDelay=2000&fileSystemType=LOCAL&chunkSize=5").to("mock:result");
+                from("direct:start").to("hdfs:localhost/" + file.toUri() + "?fileSystemType=LOCAL&splitStrategy=BYTES:5,IDLE:1000");
+                from("hdfs:localhost/" + file.toUri() + "?initialDelay=2000&fileSystemType=LOCAL&chunkSize=5").to("mock:result");
             }
         });
         context.start();
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerSplitTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerSplitTest.java
index cead95d..af9a338 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerSplitTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerSplitTest.java
@@ -125,11 +125,11 @@ public class HdfsProducerSplitTest extends HdfsTestSupport {
         return new RouteBuilder() {
             @Override
             public void configure() throws Exception {
-                from("direct:start1").to("hdfs2:localhost/" + BASE_FILE.toUri() + "1?fileSystemType=LOCAL&splitStrategy=MESSAGES:1");
-                from("direct:start2").to("hdfs2:localhost/" + BASE_FILE.toUri() + "2?fileSystemType=LOCAL&splitStrategy=BYTES:5");
-                from("direct:start3").to("hdfs2:localhost/" + BASE_FILE.toUri() + "3?fileSystemType=LOCAL&splitStrategy=IDLE:1000");
-                from("direct:start4").to("hdfs2:localhost/" + BASE_FILE.toUri() + "4?fileSystemType=LOCAL&splitStrategy=IDLE:1000,MESSAGES:1");
-                from("direct:start5").to("hdfs2:localhost/" + BASE_FILE.toUri() + "5?fileSystemType=LOCAL&splitStrategy=IDLE:1000,BYTES:5");
+                from("direct:start1").to("hdfs:localhost/" + BASE_FILE.toUri() + "1?fileSystemType=LOCAL&splitStrategy=MESSAGES:1");
+                from("direct:start2").to("hdfs:localhost/" + BASE_FILE.toUri() + "2?fileSystemType=LOCAL&splitStrategy=BYTES:5");
+                from("direct:start3").to("hdfs:localhost/" + BASE_FILE.toUri() + "3?fileSystemType=LOCAL&splitStrategy=IDLE:1000");
+                from("direct:start4").to("hdfs:localhost/" + BASE_FILE.toUri() + "4?fileSystemType=LOCAL&splitStrategy=IDLE:1000,MESSAGES:1");
+                from("direct:start5").to("hdfs:localhost/" + BASE_FILE.toUri() + "5?fileSystemType=LOCAL&splitStrategy=IDLE:1000,BYTES:5");
             }
         };
     }
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerTest.java
index 95d52ed..c6814c3 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/HdfsProducerTest.java
@@ -390,35 +390,35 @@ public class HdfsProducerTest extends HdfsTestSupport {
 
             @Override
             public void configure() throws Exception {
-                from("direct:start1").to("hdfs2://localhost/" + TEMP_DIR.toUri() + "/test-camel1?fileSystemType=LOCAL&valueType=TEXT&fileType=SEQUENCE_FILE");
+                from("direct:start1").to("hdfs://localhost/" + TEMP_DIR.toUri() + "/test-camel1?fileSystemType=LOCAL&valueType=TEXT&fileType=SEQUENCE_FILE");
 
                 /* For testing writables */
-                from("direct:write_boolean").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-boolean?fileSystemType=LOCAL&valueType=BOOLEAN&fileType=SEQUENCE_FILE");
+                from("direct:write_boolean").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-boolean?fileSystemType=LOCAL&valueType=BOOLEAN&fileType=SEQUENCE_FILE");
 
-                from("direct:write_byte").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-byte?fileSystemType=LOCAL&valueType=BYTE&fileType=SEQUENCE_FILE");
+                from("direct:write_byte").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-byte?fileSystemType=LOCAL&valueType=BYTE&fileType=SEQUENCE_FILE");
 
-                from("direct:write_int").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-int?fileSystemType=LOCAL&valueType=INT&fileType=SEQUENCE_FILE");
+                from("direct:write_int").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-int?fileSystemType=LOCAL&valueType=INT&fileType=SEQUENCE_FILE");
 
-                from("direct:write_float").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-float?fileSystemType=LOCAL&valueType=FLOAT&fileType=SEQUENCE_FILE");
+                from("direct:write_float").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-float?fileSystemType=LOCAL&valueType=FLOAT&fileType=SEQUENCE_FILE");
 
-                from("direct:write_long").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-long?fileSystemType=LOCAL&valueType=LONG&fileType=SEQUENCE_FILE");
+                from("direct:write_long").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-long?fileSystemType=LOCAL&valueType=LONG&fileType=SEQUENCE_FILE");
 
-                from("direct:write_double").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-double?fileSystemType=LOCAL&valueType=DOUBLE&fileType=SEQUENCE_FILE");
+                from("direct:write_double").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-double?fileSystemType=LOCAL&valueType=DOUBLE&fileType=SEQUENCE_FILE");
 
-                from("direct:write_text1").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-text1?fileSystemType=LOCAL&valueType=TEXT&fileType=SEQUENCE_FILE");
+                from("direct:write_text1").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-text1?fileSystemType=LOCAL&valueType=TEXT&fileType=SEQUENCE_FILE");
 
                 /* For testing key and value writing */
-                from("direct:write_text2").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-text2?fileSystemType=LOCAL&keyType=TEXT&valueType=TEXT&fileType=SEQUENCE_FILE");
+                from("direct:write_text2").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-text2?fileSystemType=LOCAL&keyType=TEXT&valueType=TEXT&fileType=SEQUENCE_FILE");
 
-                from("direct:write_text3").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-text3?fileSystemType=LOCAL&keyType=TEXT&valueType=TEXT&fileType=MAP_FILE");
+                from("direct:write_text3").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-text3?fileSystemType=LOCAL&keyType=TEXT&valueType=TEXT&fileType=MAP_FILE");
 
                 /* For testing ArrayFile */
-                from("direct:write_text4").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-text4?fileSystemType=LOCAL&valueType=TEXT&fileType=ARRAY_FILE");
+                from("direct:write_text4").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-text4?fileSystemType=LOCAL&valueType=TEXT&fileType=ARRAY_FILE");
 
                 /* For testing BloomMapFile */
-                from("direct:write_text5").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-text5?fileSystemType=LOCAL&keyType=TEXT&valueType=TEXT&fileType=BLOOMMAP_FILE");
+                from("direct:write_text5").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-text5?fileSystemType=LOCAL&keyType=TEXT&valueType=TEXT&fileType=BLOOMMAP_FILE");
 
-                from("direct:write_dynamic_filename").to("hdfs2:localhost/" + TEMP_DIR.toUri() + "/test-camel-dynamic/?fileSystemType=LOCAL&valueType=TEXT");
+                from("direct:write_dynamic_filename").to("hdfs:localhost/" + TEMP_DIR.toUri() + "/test-camel-dynamic/?fileSystemType=LOCAL&valueType=TEXT");
             }
         };
     }
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsAppendTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsAppendTest.java
index 7e242c2..300416a 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsAppendTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsAppendTest.java
@@ -61,7 +61,7 @@ public class HdfsAppendTest extends CamelTestSupport {
         context.addRoutes(new RouteBuilder() {
             @Override
             public void configure() throws Exception {
-                from("direct:start1").to("hdfs2://localhost:9000/tmp/test/test-camel-simple-write-file1?append=true&fileSystemType=HDFS");
+                from("direct:start1").to("hdfs://localhost:9000/tmp/test/test-camel-simple-write-file1?append=true&fileSystemType=HDFS");
             }
         });
         startCamelContext();
@@ -91,17 +91,17 @@ public class HdfsAppendTest extends CamelTestSupport {
         context.addRoutes(new RouteBuilder() {
             @Override
             public void configure() throws Exception {
-                from("direct:start1").to("hdfs2://localhost:9000/tmp/test-dynamic/?append=true&fileSystemType=HDFS");
+                from("direct:start1").to("hdfs://localhost:9000/tmp/test-dynamic/?append=true&fileSystemType=HDFS");
             }
         });
         startCamelContext();
 
         for (int i = 0; i < ITERATIONS; ++i) {
-            template.sendBodyAndHeader("direct:start1", "HELLO", Exchange.FILE_NAME, "camel-hdfs2.log");
+            template.sendBodyAndHeader("direct:start1", "HELLO", Exchange.FILE_NAME, "camel-hdfs.log");
         }
 
         Configuration conf = new Configuration();
-        Path file = new Path("hdfs://localhost:9000/tmp/test-dynamic/camel-hdfs2.log");
+        Path file = new Path("hdfs://localhost:9000/tmp/test-dynamic/camel-hdfs.log");
         FileSystem fs = FileSystem.get(file.toUri(), conf);
         FSDataInputStream in = fs.open(file);
         byte[] buffer = new byte[5];
diff --git a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsProducerConsumerIntegrationTest.java b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsProducerConsumerIntegrationTest.java
index d7db231..4356ec3 100644
--- a/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsProducerConsumerIntegrationTest.java
+++ b/components/camel-hdfs/src/test/java/org/apache/camel/component/hdfs2/integration/HdfsProducerConsumerIntegrationTest.java
@@ -52,8 +52,8 @@ public class HdfsProducerConsumerIntegrationTest extends CamelTestSupport {
         context.addRoutes(new RouteBuilder() {
             @Override
             public void configure() {
-                from("direct:start").to("hdfs2://localhost:9000/tmp/test/test-camel-simple-write-file?fileSystemType=HDFS&splitStrategy=BYTES:5,IDLE:1000");
-                from("hdfs2://localhost:9000/tmp/test/test-camel-simple-write-file?pattern=*&initialDelay=2000&fileSystemType=HDFS&chunkSize=5").to("mock:result");
+                from("direct:start").to("hdfs://localhost:9000/tmp/test/test-camel-simple-write-file?fileSystemType=HDFS&splitStrategy=BYTES:5,IDLE:1000");
+                from("hdfs://localhost:9000/tmp/test/test-camel-simple-write-file?pattern=*&initialDelay=2000&fileSystemType=HDFS&chunkSize=5").to("mock:result");
             }
         });
         context.start();
@@ -107,10 +107,10 @@ public class HdfsProducerConsumerIntegrationTest extends CamelTestSupport {
             @Override
             public void configure() {
                 // difference in chunkSize only to allow multiple consumers
-                from("hdfs2://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=128").to("mock:result");
-                from("hdfs2://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=256").to("mock:result");
-                from("hdfs2://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=512").to("mock:result");
-                from("hdfs2://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=1024").to("mock:result");
+                from("hdfs://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=128").to("mock:result");
+                from("hdfs://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=256").to("mock:result");
+                from("hdfs://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=512").to("mock:result");
+                from("hdfs://localhost:9000/tmp/test/multiple-consumers?pattern=*.txt&fileSystemType=HDFS&chunkSize=1024").to("mock:result");
             }
         });
         context.start();
diff --git a/components/camel-hdfs/src/test/resources/log4j2.properties b/components/camel-hdfs/src/test/resources/log4j2.properties
index bc0c8a1..76695b7 100644
--- a/components/camel-hdfs/src/test/resources/log4j2.properties
+++ b/components/camel-hdfs/src/test/resources/log4j2.properties
@@ -17,7 +17,7 @@
 
 appender.file.type = File
 appender.file.name = file
-appender.file.fileName = target/camel-hdfs2-test.log
+appender.file.fileName = target/camel-hdfs-test.log
 appender.file.layout.type = PatternLayout
 appender.file.layout.pattern = %d [%-15.15t] %-5p %-30.30c{1} - %m%n
 appender.out.type = Console
diff --git a/components/pom.xml b/components/pom.xml
index 8231abe..b6c2da4 100644
--- a/components/pom.xml
+++ b/components/pom.xml
@@ -186,7 +186,7 @@
         <module>camel-guava-eventbus</module>
         <module>camel-hazelcast</module>
         <module>camel-hbase</module>
-        <module>camel-hdfs2</module>
+        <module>camel-hdfs</module>
         <module>camel-hipchat</module>
         <module>camel-hl7</module>
         <module>camel-ical</module>
diff --git a/core/camel-endpointdsl/src/main/java/org/apache/camel/builder/endpoint/dsl/HdfsEndpointBuilderFactory.java b/core/camel-endpointdsl/src/main/java/org/apache/camel/builder/endpoint/dsl/HdfsEndpointBuilderFactory.java
index 7439d3c..95b894c 100644
--- a/core/camel-endpointdsl/src/main/java/org/apache/camel/builder/endpoint/dsl/HdfsEndpointBuilderFactory.java
+++ b/core/camel-endpointdsl/src/main/java/org/apache/camel/builder/endpoint/dsl/HdfsEndpointBuilderFactory.java
@@ -39,7 +39,7 @@ public interface HdfsEndpointBuilderFactory {
 
 
     /**
-     * Builder for endpoint consumers for the HDFS2 component.
+     * Builder for endpoint consumers for the HDFS component.
      */
     public interface HdfsEndpointConsumerBuilder
             extends
@@ -617,7 +617,7 @@ public interface HdfsEndpointBuilderFactory {
     }
 
     /**
-     * Advanced builder for endpoint consumers for the HDFS2 component.
+     * Advanced builder for endpoint consumers for the HDFS component.
      */
     public interface AdvancedHdfsEndpointConsumerBuilder
             extends
@@ -991,7 +991,7 @@ public interface HdfsEndpointBuilderFactory {
     }
 
     /**
-     * Builder for endpoint producers for the HDFS2 component.
+     * Builder for endpoint producers for the HDFS component.
      */
     public interface HdfsEndpointProducerBuilder
             extends
@@ -1235,7 +1235,7 @@ public interface HdfsEndpointBuilderFactory {
     }
 
     /**
-     * Advanced builder for endpoint producers for the HDFS2 component.
+     * Advanced builder for endpoint producers for the HDFS component.
      */
     public interface AdvancedHdfsEndpointProducerBuilder
             extends
@@ -1520,7 +1520,7 @@ public interface HdfsEndpointBuilderFactory {
     }
 
     /**
-     * Builder for endpoint for the HDFS2 component.
+     * Builder for endpoint for the HDFS component.
      */
     public interface HdfsEndpointBuilder
             extends
@@ -1676,7 +1676,7 @@ public interface HdfsEndpointBuilderFactory {
     }
 
     /**
-     * Advanced builder for endpoint for the HDFS2 component.
+     * Advanced builder for endpoint for the HDFS component.
      */
     public interface AdvancedHdfsEndpointBuilder
             extends
@@ -2011,7 +2011,7 @@ public interface HdfsEndpointBuilderFactory {
         BLOCK;
     }
     /**
-     * HDFS2 (camel-hdfs)
+     * HDFS (camel-hdfs)
      * For reading/writing from/to an HDFS filesystem using Hadoop 2.x.
      * 
      * Category: hadoop,file
diff --git a/parent/pom.xml b/parent/pom.xml
index 38cc3ff..9bb17eb 100644
--- a/parent/pom.xml
+++ b/parent/pom.xml
@@ -1389,7 +1389,7 @@
       </dependency>
       <dependency>
         <groupId>org.apache.camel</groupId>
-        <artifactId>camel-hdfs2</artifactId>
+        <artifactId>camel-hdfs</artifactId>
         <version>${project.version}</version>
       </dependency>
       <dependency>
@@ -2933,7 +2933,7 @@
       </dependency>
       <dependency>
         <groupId>org.apache.camel</groupId>
-        <artifactId>camel-hdfs2-starter</artifactId>
+        <artifactId>camel-hdfs-starter</artifactId>
         <version>${project.version}</version>
       </dependency>
       <dependency>
diff --git a/platforms/spring-boot/components-starter/camel-hdfs-starter/src/main/java/org/apache/camel/component/hdfs2/springboot/HdfsComponentAutoConfiguration.java b/platforms/spring-boot/components-starter/camel-hdfs-starter/src/main/java/org/apache/camel/component/hdfs2/springboot/HdfsComponentAutoConfiguration.java
index 34c1f3e..829973c 100644
--- a/platforms/spring-boot/components-starter/camel-hdfs-starter/src/main/java/org/apache/camel/component/hdfs2/springboot/HdfsComponentAutoConfiguration.java
+++ b/platforms/spring-boot/components-starter/camel-hdfs-starter/src/main/java/org/apache/camel/component/hdfs2/springboot/HdfsComponentAutoConfiguration.java
@@ -75,7 +75,7 @@ public class HdfsComponentAutoConfiguration {
     }
 
     @Lazy
-    @Bean({"hdfs-component", "hdfs2-component"})
+    @Bean(name = "hdfs-component")
     @ConditionalOnMissingBean(HdfsComponent.class)
     public HdfsComponent configureHdfsComponent() throws Exception {
         HdfsComponent component = new HdfsComponent();
diff --git a/platforms/spring-boot/components-starter/pom.xml b/platforms/spring-boot/components-starter/pom.xml
index 3ffdd06..e4887f7 100644
--- a/platforms/spring-boot/components-starter/pom.xml
+++ b/platforms/spring-boot/components-starter/pom.xml
@@ -210,7 +210,6 @@
     <module>camel-hazelcast-starter</module>
     <module>camel-hbase-starter</module>
     <module>camel-hdfs-starter</module>
-    <module>camel-hdfs2-starter</module>
     <module>camel-hipchat-starter</module>
     <module>camel-hl7-starter</module>
     <module>camel-http-common-starter</module>