You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by echarles <gi...@git.apache.org> on 2018/05/04 09:36:28 UTC

[GitHub] spark issue #20451: [SPARK-23146][WIP] Support client mode for Kubernetes cl...

Github user echarles commented on the issue:

    https://github.com/apache/spark/pull/20451
  
    I have updated this branch and tested successfully client mode OutCluster.
    
    Happy to get confirmation of this from anyone here (cc/ @foxish)
    
    For InCluster (was working fine before ##20910), worker starts but fails due to UnknownHostException: The worker can not resolve the Driver hostname): @mccheah was there anything changed in terms of HeadlessDriver step during the #20910 refactoring?
    
    ```
    k logs -f spark-exec-1
    ++ id -u
    + myuid=0
    ++ id -g
    + mygid=0
    + set +e
    ++ getent passwd 0
    + uidentry=root:x:0:0:root:/root:/bin/ash
    + set -e
    + '[' -z root:x:0:0:root:/root:/bin/ash ']'
    + SPARK_K8S_CMD=executor
    + '[' -z executor ']'
    + shift 1
    + SPARK_CLASSPATH=':/opt/spark/jars/*'
    + env
    + grep SPARK_JAVA_OPT_
    + sort -t_ -k4 -n
    + sed 's/[^=]*=\(.*\)/\1/g'
    + readarray -t SPARK_JAVA_OPTS
    + '[' -n '' ']'
    + '[' -n '' ']'
    + case "$SPARK_K8S_CMD" in
    + CMD=(${JAVA_HOME}/bin/java "${SPARK_JAVA_OPTS[@]}" -Xms$SPARK_EXECUTOR_MEMORY -Xmx$SPARK_EXECUTOR_MEMORY -cp "$SPARK_CLASSPATH" org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url $SPARK_DRIVER_URL --executor-id $SPARK_EXECUTOR_ID --cores $SPARK_EXECUTOR_CORES --app-id $SPARK_APPLICATION_ID --hostname $SPARK_EXECUTOR_POD_IP)
    + exec /sbin/tini -s -- /usr/lib/jvm/java-1.8-openjdk/bin/java -Xms1g -Xmx1g -cp ':/opt/spark/jars/*' org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url spark://CoarseGrainedScheduler@spark-pod:33913 --executor-id 1 --cores 1 --app-id spark-application-1525426093472 --hostname 172.17.0.9
    2018-05-04 09:28:16 INFO  CoarseGrainedExecutorBackend:2502 - Started daemon with process name: 15@spark-exec-1
    2018-05-04 09:28:16 INFO  SignalUtils:54 - Registered signal handler for TERM
    2018-05-04 09:28:16 INFO  SignalUtils:54 - Registered signal handler for HUP
    2018-05-04 09:28:16 INFO  SignalUtils:54 - Registered signal handler for INT
    2018-05-04 09:28:17 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
    2018-05-04 09:28:17 INFO  SecurityManager:54 - Changing view acls to: root
    2018-05-04 09:28:17 INFO  SecurityManager:54 - Changing modify acls to: root
    2018-05-04 09:28:17 INFO  SecurityManager:54 - Changing view acls groups to: 
    2018-05-04 09:28:17 INFO  SecurityManager:54 - Changing modify acls groups to: 
    2018-05-04 09:28:17 INFO  SecurityManager:54 - SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(root); groups with view permissions: Set(); users  with modify permissions: Set(root); groups with modify permissions: Set()
    Exception in thread "main" java.lang.reflect.UndeclaredThrowableException
    	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1904)
    	at org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:65)
    	at org.apache.spark.executor.CoarseGrainedExecutorBackend$.run(CoarseGrainedExecutorBackend.scala:188)
    	at org.apache.spark.executor.CoarseGrainedExecutorBackend$.main(CoarseGrainedExecutorBackend.scala:281)
    	at org.apache.spark.executor.CoarseGrainedExecutorBackend.main(CoarseGrainedExecutorBackend.scala)
    Caused by: org.apache.spark.SparkException: Exception thrown in awaitResult: 
    	at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:205)
    	at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
    	at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)
    	at org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$run$1.apply$mcV$sp(CoarseGrainedExecutorBackend.scala:201)
    	at org.apache.spark.deploy.SparkHadoopUtil$$anon$2.run(SparkHadoopUtil.scala:66)
    	at org.apache.spark.deploy.SparkHadoopUtil$$anon$2.run(SparkHadoopUtil.scala:65)
    	at java.security.AccessController.doPrivileged(Native Method)
    	at javax.security.auth.Subject.doAs(Subject.java:422)
    	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
    	... 4 more
    Caused by: java.io.IOException: Failed to connect to spark-pod:33913
    	at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:245)
    	at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:187)
    	at org.apache.spark.rpc.netty.NettyRpcEnv.createClient(NettyRpcEnv.scala:198)
    	at org.apache.spark.rpc.netty.Outbox$$anon$1.call(Outbox.scala:194)
    	at org.apache.spark.rpc.netty.Outbox$$anon$1.call(Outbox.scala:190)
    	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    	at java.lang.Thread.run(Thread.java:748)
    Caused by: java.net.UnknownHostException: spark-pod
    	at java.net.InetAddress.getAllByName0(InetAddress.java:1280)
    	at java.net.InetAddress.getAllByName(InetAddress.java:1192)
    	at java.net.InetAddress.getAllByName(InetAddress.java:1126)
    	at java.net.InetAddress.getByName(InetAddress.java:1076)
    	at io.netty.util.internal.SocketUtils$8.run(SocketUtils.java:146)
    	at io.netty.util.internal.SocketUtils$8.run(SocketUtils.java:143)
    	at java.security.AccessController.doPrivileged(Native Method)
    	at io.netty.util.internal.SocketUtils.addressByName(SocketUtils.java:143)
    	at io.netty.resolver.DefaultNameResolver.doResolve(DefaultNameResolver.java:43)
    	at io.netty.resolver.SimpleNameResolver.resolve(SimpleNameResolver.java:63)
    	at io.netty.resolver.SimpleNameResolver.resolve(SimpleNameResolver.java:55)
    	at io.netty.resolver.InetSocketAddressResolver.doResolve(InetSocketAddressResolver.java:57)
    	at io.netty.resolver.InetSocketAddressResolver.doResolve(InetSocketAddressResolver.java:32)
    	at io.netty.resolver.AbstractAddressResolver.resolve(AbstractAddressResolver.java:108)
    	at io.netty.bootstrap.Bootstrap.doResolveAndConnect0(Bootstrap.java:208)
    	at io.netty.bootstrap.Bootstrap.access$000(Bootstrap.java:49)
    	at io.netty.bootstrap.Bootstrap$1.operationComplete(Bootstrap.java:188)
    	at io.netty.bootstrap.Bootstrap$1.operationComplete(Bootstrap.java:174)
    	at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:507)
    	at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:481)
    	at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:420)
    	at io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:104)
    	at io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:82)
    	at io.netty.channel.AbstractChannel$AbstractUnsafe.safeSetSuccess(AbstractChannel.java:978)
    	at io.netty.channel.AbstractChannel$AbstractUnsafe.register0(AbstractChannel.java:512)
    	at io.netty.channel.AbstractChannel$AbstractUnsafe.access$200(AbstractChannel.java:423)
    	at io.netty.channel.AbstractChannel$AbstractUnsafe$1.run(AbstractChannel.java:482)
    	at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
    	at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:403)
    	at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:463)
    	at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:858)
    	at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
    	... 1 more
    ```


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org