You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dongjoon Hyun (Jira)" <ji...@apache.org> on 2020/01/13 05:24:00 UTC
[jira] [Resolved] (SPARK-30204) Support for config Pod DNS for
Kubernetes
[ https://issues.apache.org/jira/browse/SPARK-30204?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Dongjoon Hyun resolved SPARK-30204.
-----------------------------------
Resolution: Workaround
This PR is not merged because we can use POD template.
> Support for config Pod DNS for Kubernetes
> -----------------------------------------
>
> Key: SPARK-30204
> URL: https://issues.apache.org/jira/browse/SPARK-30204
> Project: Spark
> Issue Type: Improvement
> Components: Kubernetes
> Affects Versions: 3.0.0
> Reporter: vanderliang
> Priority: Major
>
> Current we can not configure the pod dns nameservers and searches when submit a job via cli for kubernetes. However, this's a common scenarios for hybricloud where we use public cloud compute resourses while with private dns.
>
> {code:java}
> //代码占位符
> apiVersion: v1
> kind: Pod
> metadata:
> namespace: default
> name: dns-example
> spec:
> containers:
> - name: test
> image: nginx
> dnsConfig:
> nameservers:
> - 1.2.3.4
> searches:
> - ns1.svc.cluster-domain.example
> - my.dns.search.suffix
> options:
> - name: ndots
> value: "2"
> - name: edns0
> {code}
> As a result, we can use the following property to specify the pod dns config.
> * spark.kubernetes.dnsConfig.nameservers, Comma separated list of the Kubernetes dns nameservers for driver and executor.
> * spark.kubernetes.dnsConfig.searches, Comma separated list of the Kubernetes dns searches for driver and executor.
> * spark.kubernetes.dnsConfig.options.[OptionVariableName], Add the dns option variable specified by OptionVariableName to the Driver And Executor process. The user can specify multiple of these to set multiple options variables.
>
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org