You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@seatunnel.apache.org by fa...@apache.org on 2023/01/16 07:11:35 UTC

[incubator-seatunnel] branch dev updated: [Doc] [Chore] Remove v1 doc in config directory (#3955)

This is an automated email from the ASF dual-hosted git repository.

fanjia pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/incubator-seatunnel.git


The following commit(s) were added to refs/heads/dev by this push:
     new c3b929d39 [Doc] [Chore] Remove v1 doc in config directory (#3955)
c3b929d39 is described below

commit c3b929d3925b070d2685897b1ba980aeb1107fea
Author: Guangdong Liu <80...@qq.com>
AuthorDate: Mon Jan 16 15:11:28 2023 +0800

    [Doc] [Chore] Remove v1 doc in config directory (#3955)
---
 config/flink.sql.conf.template       | 49 -----------------------
 config/spark.batch.conf.template     | 72 ---------------------------------
 config/spark.streaming.conf.template | 77 ------------------------------------
 3 files changed, 198 deletions(-)

diff --git a/config/flink.sql.conf.template b/config/flink.sql.conf.template
deleted file mode 100644
index 4742a04fb..000000000
--- a/config/flink.sql.conf.template
+++ /dev/null
@@ -1,49 +0,0 @@
---
--- Licensed to the Apache Software Foundation (ASF) under one or more
--- contributor license agreements.  See the NOTICE file distributed with
--- this work for additional information regarding copyright ownership.
--- The ASF licenses this file to You under the Apache License, Version 2.0
--- (the "License"); you may not use this file except in compliance with
--- the License.  You may obtain a copy of the License at
---
---     http://www.apache.org/licenses/LICENSE-2.0
---
--- Unless required by applicable law or agreed to in writing, software
--- distributed under the License is distributed on an "AS IS" BASIS,
--- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
--- See the License for the specific language governing permissions and
--- limitations under the License.
---
-
---
--- This config file is a demonstration of sql processing in SeaTunnel config
---
---
-
-
-SET table.dml-sync = true;
-
-CREATE TABLE events (
-  f_type INT,
-  f_uid INT,
-  ts AS localtimestamp,
-  WATERMARK FOR ts AS ts
-) WITH (
-  'connector' = 'datagen',
-  'rows-per-second'='5',
-  'fields.f_type.min'='1',
-  'fields.f_type.max'='5',
-  'fields.f_uid.min'='1',
-  'fields.f_uid.max'='1000'
-);
-
-CREATE TABLE print_table (
-  type INT,
-  uid INT,
-  lstmt TIMESTAMP
-) WITH (
-  'connector' = 'print',
-  'sink.parallelism' = '1'
-);
-
-INSERT INTO print_table SELECT * FROM events where f_type = 1;
diff --git a/config/spark.batch.conf.template b/config/spark.batch.conf.template
deleted file mode 100644
index 674609d4b..000000000
--- a/config/spark.batch.conf.template
+++ /dev/null
@@ -1,72 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-######
-###### This config file is a demonstration of batch processing in SeaTunnel config
-######
-
-env {
-  # You can set spark configuration here
-  # see available properties defined by spark: https://spark.apache.org/docs/latest/configuration.html#available-properties
-  spark.app.name = "SeaTunnel"
-  spark.executor.instances = 2
-  spark.executor.cores = 1
-  spark.executor.memory = "1g"
-}
-
-source {
-  # This is a example input plugin **only for test and demonstrate the feature input plugin**
-  Fake {
-    result_table_name = "my_dataset"
-  }
-
-  # You can also use other input plugins, such as file
-  # file {
-  #   result_table_name = "accesslog"
-  #   path = "hdfs://hadoop-cluster-01/nginx/accesslog"
-  #   format = "json"
-  # }
-
-  # If you would like to get more information about how to configure seatunnel and see full list of input plugins,
-  # please go to https://seatunnel.apache.org/docs/spark/configuration/source-plugins/Fake
-}
-
-transform {
-  # split data by specific delimiter
-
-  # you can also use other filter plugins, such as sql
-  # sql {
-  #   sql = "select * from accesslog where request_time > 1000"
-  # }
-
-  # If you would like to get more information about how to configure seatunnel and see full list of filter plugins,
-  # please go to https://seatunnel.apache.org/docs/spark/configuration/transform-plugins/Sql
-}
-
-sink {
-  # choose stdout output plugin to output data to console
-  Console {}
-
-  # you can also use other output plugins, such as hdfs
-  # hdfs {
-  #   path = "hdfs://hadoop-cluster-01/nginx/accesslog_processed"
-  #   save_mode = "append"
-  # }
-
-  # If you would like to get more information about how to configure seatunnel and see full list of output plugins,
-  # please go to https://seatunnel.apache.org/docs/spark/configuration/sink-plugins/Console
-}
diff --git a/config/spark.streaming.conf.template b/config/spark.streaming.conf.template
deleted file mode 100644
index 2b09071b6..000000000
--- a/config/spark.streaming.conf.template
+++ /dev/null
@@ -1,77 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-######
-###### This config file is a demonstration of stream processing in seatunnel config
-######
-
-env {
-  # You can set spark configuration here
-  # see available properties defined by spark: https://spark.apache.org/docs/latest/configuration.html#available-properties
-  spark.app.name = "SeaTunnel"
-  spark.executor.instances = 2
-  spark.executor.cores = 1
-  spark.executor.memory = "1g"
-  spark.stream.batchDuration = 5
-}
-
-source {
-  # This is a example input plugin **only for test and demonstrate the feature input plugin**
-  FakeStream {
-    content = ["Hello World, SeaTunnel"]
-  }
-
-  # You can also use other input plugins, such as file
-  # file {
-  #   result_table_name = "accesslog"
-  #   path = "hdfs://hadoop-cluster-01/nginx/accesslog"
-  #   format = "json"
-  # }
-
-  # If you would like to get more information about how to configure seatunnel and see full list of input plugins,
-  # please go to https://seatunnel.apache.org/docs/spark/configuration/source-plugins/FakeStream
-}
-
-transform {
-
-  split {
-    fields = ["msg", "name"]
-    delimiter = ","
-  }
-
-  # you can also use other filter plugins, such as sql
-  # Sql {
-  #   sql = "select * from accesslog where request_time > 1000"
-  # }
-
-  # If you would like to get more information about how to configure seatunnel and see full list of filter plugins,
-  # please go to https://seatunnel.apache.org/docs/spark/configuration/transform-plugins/Split
-}
-
-sink {
-  # choose stdout output plugin to output data to console
-  Console {}
-
-  # you can also use other output plugins, such as hdfs
-  # hdfs {
-  #   path = "hdfs://hadoop-cluster-01/nginx/accesslog_processed"
-  #   save_mode = "append"
-  # }
-
-  # If you would like to get more information about how to configure seatunnel and see full list of output plugins,
-  # please go to https://seatunnel.apache.org/docs/spark/configuration/sink-plugins/Console
-}