You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@seatunnel.apache.org by fa...@apache.org on 2023/01/16 07:11:35 UTC
[incubator-seatunnel] branch dev updated: [Doc] [Chore] Remove v1 doc in config directory (#3955)
This is an automated email from the ASF dual-hosted git repository.
fanjia pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/incubator-seatunnel.git
The following commit(s) were added to refs/heads/dev by this push:
new c3b929d39 [Doc] [Chore] Remove v1 doc in config directory (#3955)
c3b929d39 is described below
commit c3b929d3925b070d2685897b1ba980aeb1107fea
Author: Guangdong Liu <80...@qq.com>
AuthorDate: Mon Jan 16 15:11:28 2023 +0800
[Doc] [Chore] Remove v1 doc in config directory (#3955)
---
config/flink.sql.conf.template | 49 -----------------------
config/spark.batch.conf.template | 72 ---------------------------------
config/spark.streaming.conf.template | 77 ------------------------------------
3 files changed, 198 deletions(-)
diff --git a/config/flink.sql.conf.template b/config/flink.sql.conf.template
deleted file mode 100644
index 4742a04fb..000000000
--- a/config/flink.sql.conf.template
+++ /dev/null
@@ -1,49 +0,0 @@
---
--- Licensed to the Apache Software Foundation (ASF) under one or more
--- contributor license agreements. See the NOTICE file distributed with
--- this work for additional information regarding copyright ownership.
--- The ASF licenses this file to You under the Apache License, Version 2.0
--- (the "License"); you may not use this file except in compliance with
--- the License. You may obtain a copy of the License at
---
--- http://www.apache.org/licenses/LICENSE-2.0
---
--- Unless required by applicable law or agreed to in writing, software
--- distributed under the License is distributed on an "AS IS" BASIS,
--- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
--- See the License for the specific language governing permissions and
--- limitations under the License.
---
-
---
--- This config file is a demonstration of sql processing in SeaTunnel config
---
---
-
-
-SET table.dml-sync = true;
-
-CREATE TABLE events (
- f_type INT,
- f_uid INT,
- ts AS localtimestamp,
- WATERMARK FOR ts AS ts
-) WITH (
- 'connector' = 'datagen',
- 'rows-per-second'='5',
- 'fields.f_type.min'='1',
- 'fields.f_type.max'='5',
- 'fields.f_uid.min'='1',
- 'fields.f_uid.max'='1000'
-);
-
-CREATE TABLE print_table (
- type INT,
- uid INT,
- lstmt TIMESTAMP
-) WITH (
- 'connector' = 'print',
- 'sink.parallelism' = '1'
-);
-
-INSERT INTO print_table SELECT * FROM events where f_type = 1;
diff --git a/config/spark.batch.conf.template b/config/spark.batch.conf.template
deleted file mode 100644
index 674609d4b..000000000
--- a/config/spark.batch.conf.template
+++ /dev/null
@@ -1,72 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements. See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-######
-###### This config file is a demonstration of batch processing in SeaTunnel config
-######
-
-env {
- # You can set spark configuration here
- # see available properties defined by spark: https://spark.apache.org/docs/latest/configuration.html#available-properties
- spark.app.name = "SeaTunnel"
- spark.executor.instances = 2
- spark.executor.cores = 1
- spark.executor.memory = "1g"
-}
-
-source {
- # This is a example input plugin **only for test and demonstrate the feature input plugin**
- Fake {
- result_table_name = "my_dataset"
- }
-
- # You can also use other input plugins, such as file
- # file {
- # result_table_name = "accesslog"
- # path = "hdfs://hadoop-cluster-01/nginx/accesslog"
- # format = "json"
- # }
-
- # If you would like to get more information about how to configure seatunnel and see full list of input plugins,
- # please go to https://seatunnel.apache.org/docs/spark/configuration/source-plugins/Fake
-}
-
-transform {
- # split data by specific delimiter
-
- # you can also use other filter plugins, such as sql
- # sql {
- # sql = "select * from accesslog where request_time > 1000"
- # }
-
- # If you would like to get more information about how to configure seatunnel and see full list of filter plugins,
- # please go to https://seatunnel.apache.org/docs/spark/configuration/transform-plugins/Sql
-}
-
-sink {
- # choose stdout output plugin to output data to console
- Console {}
-
- # you can also use other output plugins, such as hdfs
- # hdfs {
- # path = "hdfs://hadoop-cluster-01/nginx/accesslog_processed"
- # save_mode = "append"
- # }
-
- # If you would like to get more information about how to configure seatunnel and see full list of output plugins,
- # please go to https://seatunnel.apache.org/docs/spark/configuration/sink-plugins/Console
-}
diff --git a/config/spark.streaming.conf.template b/config/spark.streaming.conf.template
deleted file mode 100644
index 2b09071b6..000000000
--- a/config/spark.streaming.conf.template
+++ /dev/null
@@ -1,77 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements. See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-######
-###### This config file is a demonstration of stream processing in seatunnel config
-######
-
-env {
- # You can set spark configuration here
- # see available properties defined by spark: https://spark.apache.org/docs/latest/configuration.html#available-properties
- spark.app.name = "SeaTunnel"
- spark.executor.instances = 2
- spark.executor.cores = 1
- spark.executor.memory = "1g"
- spark.stream.batchDuration = 5
-}
-
-source {
- # This is a example input plugin **only for test and demonstrate the feature input plugin**
- FakeStream {
- content = ["Hello World, SeaTunnel"]
- }
-
- # You can also use other input plugins, such as file
- # file {
- # result_table_name = "accesslog"
- # path = "hdfs://hadoop-cluster-01/nginx/accesslog"
- # format = "json"
- # }
-
- # If you would like to get more information about how to configure seatunnel and see full list of input plugins,
- # please go to https://seatunnel.apache.org/docs/spark/configuration/source-plugins/FakeStream
-}
-
-transform {
-
- split {
- fields = ["msg", "name"]
- delimiter = ","
- }
-
- # you can also use other filter plugins, such as sql
- # Sql {
- # sql = "select * from accesslog where request_time > 1000"
- # }
-
- # If you would like to get more information about how to configure seatunnel and see full list of filter plugins,
- # please go to https://seatunnel.apache.org/docs/spark/configuration/transform-plugins/Split
-}
-
-sink {
- # choose stdout output plugin to output data to console
- Console {}
-
- # you can also use other output plugins, such as hdfs
- # hdfs {
- # path = "hdfs://hadoop-cluster-01/nginx/accesslog_processed"
- # save_mode = "append"
- # }
-
- # If you would like to get more information about how to configure seatunnel and see full list of output plugins,
- # please go to https://seatunnel.apache.org/docs/spark/configuration/sink-plugins/Console
-}