You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by ka...@apache.org on 2020/08/26 17:46:40 UTC

[airflow] branch master updated: Enhanced the Kubernetes Executor doc (#10433)

This is an automated email from the ASF dual-hosted git repository.

kaxilnaik pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/master by this push:
     new 46ac09d  Enhanced the Kubernetes Executor doc  (#10433)
46ac09d is described below

commit 46ac09d5c9b9f6e36cce0a1d3812f483ed7201eb
Author: Vikram Koka <vi...@astronomer.io>
AuthorDate: Wed Aug 26 10:42:35 2020 -0700

    Enhanced the Kubernetes Executor doc  (#10433)
    
    A simple architecture diagram to show the Airflow setup when used with the Kubernetes executor
---
 docs/executor/kubernetes.rst       |  16 +++++++++++++---
 docs/img/arch-diag-kubernetes.png  | Bin 0 -> 21804 bytes
 docs/img/arch-diag-kubernetes2.png | Bin 0 -> 86384 bytes
 3 files changed, 13 insertions(+), 3 deletions(-)

diff --git a/docs/executor/kubernetes.rst b/docs/executor/kubernetes.rst
index 3c64c16..d3664b9 100644
--- a/docs/executor/kubernetes.rst
+++ b/docs/executor/kubernetes.rst
@@ -44,15 +44,25 @@ KubernetesExecutor Architecture
 The KubernetesExecutor runs as a process in the Scheduler that only requires access to the Kubernetes API (it does *not* need to run inside of a Kubernetes cluster). The KubernetesExecutor requires a non-sqlite database in the backend, but there are no external brokers or persistent workers needed.
 For these reasons, we recommend the KubernetesExecutor for deployments have long periods of dormancy between DAG execution.
 
+When a DAG submits a task, the KubernetesExecutor requests a worker pod from the Kubernetes API. The worker pod then runs the task, reports the result, and terminates.
 
-.. image:: ../img/k8s-0-worker.jpeg
 
+.. image:: ../img/arch-diag-kubernetes.png
 
-When a DAG submits a task, the KubernetesExecutor requests a worker pod from the Kubernetes API. The worker pod then runs the task, reports the result, and terminates.
 
+In contrast to the Celery Executor, the Kubernetes Executor does not require additional components such as Redis and Flower, but does require the Kubernetes infrastructure.
+
+One example of an Airflow deployment running on a distributed set of five nodes in a Kubernetes cluster is shown below. 
+
+.. image:: ../img/arch-diag-kubernetes2.png
+
+The Kubernetes Executor has an advantage over the Celery Executor in that Pods are only spun up when required for task execution compared to the Celery Executor where the workers are statically configured and are running all the time, regardless of workloads. However, this could be a disadvantage depending on the latency needs, since a task takes longer to start using the Kubernetes Executor, since it now includes the Pod startup time.
+
+Consistent with the regular Airflow architecture, the Workers need access to the DAG files to execute the tasks within those DAGs and interact with the Metadata repository. Also, configuration information specific to the Kubernetes Executor, such as the worker namespace and image information, needs to be specified in the Airflow Configuration file.
+
+Additionally, the Kubernetes Executor enables specification of additional features on a per-task basis using the Executor config.
 
 
-.. image:: ../img/k8s-3-worker.jpeg
 
 .. @startuml
 .. Airflow_Scheduler -> Kubernetes: Request a new pod with command "airflow run..."
diff --git a/docs/img/arch-diag-kubernetes.png b/docs/img/arch-diag-kubernetes.png
new file mode 100644
index 0000000..1bbbc98
Binary files /dev/null and b/docs/img/arch-diag-kubernetes.png differ
diff --git a/docs/img/arch-diag-kubernetes2.png b/docs/img/arch-diag-kubernetes2.png
new file mode 100644
index 0000000..acaaf43
Binary files /dev/null and b/docs/img/arch-diag-kubernetes2.png differ