You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@dolphinscheduler.apache.org by GitBox <gi...@apache.org> on 2022/10/12 00:45:34 UTC

[GitHub] [dolphinscheduler] haorengoodman commented on issue #12308: Variables in spark-sql script in 3.0.1 cannot be replaced

haorengoodman commented on issue #12308:
URL: https://github.com/apache/dolphinscheduler/issues/12308#issuecomment-1275441726

   ![image](https://user-images.githubusercontent.com/6840026/195223310-5dd4d7ec-4500-458f-a7de-db58e9f09e12.png)
   
   ![image](https://user-images.githubusercontent.com/6840026/195223354-67b5457d-a033-4f87-88b1-092e7a35d66a.png)
   
   [LOG-PATH]: /opt/dolphin/install/worker-server/logs/20221012/7172773527040_35-121-137.log, [HOST]:  Host{address='10.20.20.12:1234', ip='10.20.20.12', port=1234}
   [INFO] 2022-10-12 00:42:23.980 +0000 - spark task params {"localParams":[{"prop":"dt","direct":"IN","type":"VARCHAR","value":"$[yyyy-MM-dd]"}],"rawScript":"select ${dt}","resourceList":[],"programType":"SQL","mainClass":"","deployMode":"local","sparkVersion":"SPARK2","driverCores":1,"driverMemory":"512M","numExecutors":1,"executorMemory":"1G","executorCores":1}
   [INFO] 2022-10-12 00:42:23.990 +0000 - raw script : select ${dt}
   [INFO] 2022-10-12 00:42:23.990 +0000 - task execute path : /tmp/dolphinscheduler/exec/process/7170590636544/7172773527040_35/121/137
   [INFO] 2022-10-12 00:42:24.016 +0000 - spark task command: ${SPARK_HOME2}/bin/spark-sql --master local --driver-cores 1 --driver-memory 512M --num-executors 1 --executor-cores 1 --executor-memory 1G -f /tmp/dolphinscheduler/exec/process/7170590636544/7172773527040_35/121/137/121_137_node.sql
   [INFO] 2022-10-12 00:42:24.017 +0000 - tenantCode user:hadoop, task dir:121_137
   [INFO] 2022-10-12 00:42:24.017 +0000 - create command file:/tmp/dolphinscheduler/exec/process/7170590636544/7172773527040_35/121/137/121_137.command
   [INFO] 2022-10-12 00:42:24.017 +0000 - command : #!/bin/bash
   BASEDIR=$(cd `dirname $0`; pwd)
   cd $BASEDIR
   source /opt/dolphin/install/worker-server/conf/dolphinscheduler_env.sh
   ${SPARK_HOME2}/bin/spark-sql --master local --driver-cores 1 --driver-memory 512M --num-executors 1 --executor-cores 1 --executor-memory 1G -f /tmp/dolphinscheduler/exec/process/7170590636544/7172773527040_35/121/137/121_137_node.sql
   [INFO] 2022-10-12 00:42:24.061 +0000 - task run command: sudo -u hadoop sh /tmp/dolphinscheduler/exec/process/7170590636544/7172773527040_35/121/137/121_137.command
   [INFO] 2022-10-12 00:42:24.071 +0000 - process start, process id is: 689910
   [INFO] 2022-10-12 00:42:26.063 +0000 -  -> 2022-10-12 08:42:25 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   	2022-10-12 08:42:25 INFO conf.HiveConf: Found configuration file file:/opt/usdp-srv/srv/udp/2.0.0.0/spark/conf/hive-site.xml
   	2022-10-12 08:42:25 WARN conf.HiveConf: HiveConf of name hive.enforce.bucketing does not exist
   	2022-10-12 08:42:25 INFO internal.SharedState: loading hive config file: file:/opt/usdp-srv/srv/udp/2.0.0.0/spark/conf/hive-site.xml
   	2022-10-12 08:42:25 INFO internal.SharedState: spark.sql.warehouse.dir is not set, but hive.metastore.warehouse.dir is set. Setting spark.sql.warehouse.dir to the value of hive.metastore.warehouse.dir ('/cluster1/user/hive/warehouse/').
   	2022-10-12 08:42:25 INFO internal.SharedState: Warehouse path is '/cluster1/user/hive/warehouse/'.
   	2022-10-12 08:42:25 INFO session.SessionState: Created HDFS directory: /cluster1/tmp/hive/hadoop/ab87d9d4-2d69-4865-abc8-22ff8fdf31a4
   	2022-10-12 08:42:26 INFO session.SessionState: Created local directory: /tmp/hadoop/ab87d9d4-2d69-4865-abc8-22ff8fdf31a4
   	2022-10-12 08:42:26 INFO session.SessionState: Created HDFS directory: /cluster1/tmp/hive/hadoop/ab87d9d4-2d69-4865-abc8-22ff8fdf31a4/_tmp_space.db
   [INFO] 2022-10-12 00:42:27.064 +0000 -  -> 2022-10-12 08:42:26 INFO spark.SparkContext: Running Spark version 3.0.1
   	2022-10-12 08:42:26 INFO resource.ResourceUtils: ==============================================================
   	2022-10-12 08:42:26 INFO resource.ResourceUtils: Resources for spark.driver:
   	
   	2022-10-12 08:42:26 INFO resource.ResourceUtils: ==============================================================
   	2022-10-12 08:42:26 INFO spark.SparkContext: Submitted application: SparkSQL::10.20.20.12
   	2022-10-12 08:42:26 INFO spark.SecurityManager: Changing view acls to: hadoop
   	2022-10-12 08:42:26 INFO spark.SecurityManager: Changing modify acls to: hadoop
   	2022-10-12 08:42:26 INFO spark.SecurityManager: Changing view acls groups to: 
   	2022-10-12 08:42:26 INFO spark.SecurityManager: Changing modify acls groups to: 
   	2022-10-12 08:42:26 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(hadoop); groups with view permissions: Set(); users  with modify permissions: Set(hadoop); groups with modify permissions: Set()
   	2022-10-12 08:42:26 INFO util.Utils: Successfully started service 'sparkDriver' on port 36912.
   	2022-10-12 08:42:26 INFO spark.SparkEnv: Registering MapOutputTracker
   	2022-10-12 08:42:26 INFO spark.SparkEnv: Registering BlockManagerMaster
   	2022-10-12 08:42:26 INFO storage.BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
   	2022-10-12 08:42:26 INFO storage.BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
   	2022-10-12 08:42:26 INFO spark.SparkEnv: Registering BlockManagerMasterHeartbeat
   	2022-10-12 08:42:26 INFO storage.DiskBlockManager: Created local directory at /home/data/udp/2.0.0.0/spark/blockmgr-d9518483-40a9-44c7-8af8-fba2c519508e
   	2022-10-12 08:42:26 INFO memory.MemoryStore: MemoryStore started with capacity 93.3 MiB
   	2022-10-12 08:42:26 INFO spark.SparkEnv: Registering OutputCommitCoordinator
   	2022-10-12 08:42:26 INFO util.log: Logging initialized @2161ms to org.sparkproject.jetty.util.log.Slf4jLog
   	2022-10-12 08:42:26 INFO server.Server: jetty-9.4.z-SNAPSHOT; built: 2019-04-29T20:42:08.989Z; git: e1bc35120a6617ee3df052294e433f3a25ce7097; jvm 1.8.0_202-b08
   	2022-10-12 08:42:26 INFO server.Server: Started @2227ms
   	2022-10-12 08:42:26 INFO server.AbstractConnector: Started ServerConnector@64f1fd08{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
   	2022-10-12 08:42:26 INFO util.Utils: Successfully started service 'SparkUI' on port 4040.
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@10358c32{/jobs,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5f117b3d{/jobs/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@71b6d77f{/jobs/job,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@78226c36{/jobs/job/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@11f406f8{/stages,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@622fdb81{/stages/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2ec3633f{/stages/stage,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@20576557{/stages/stage/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@45c2e0a6{/stages/pool,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@a7ad6e5{/stages/pool/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@690bc15e{/storage,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@475646d4{/storage/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3db972d2{/storage/rdd,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@687e4c93{/storage/rdd/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6ec7bce0{/environment,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3727f0ee{/environment/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@fb713e7{/executors,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@13741d5a{/executors/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5c7668ba{/executors/threadDump,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@158f4cfe{/executors/threadDump/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5aaaa446{/static,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3d37203b{/,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1894593a{/api,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5527b211{/jobs/job/kill,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@78010562{/stages/stage/kill,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at http://h12:4040
   	2022-10-12 08:42:26 INFO executor.Executor: Starting executor ID driver on host h12
   	2022-10-12 08:42:26 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 58572.
   	2022-10-12 08:42:26 INFO netty.NettyBlockTransferService: Server created on h12:58572
   	2022-10-12 08:42:26 INFO storage.BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
   	2022-10-12 08:42:26 INFO storage.BlockManagerMaster: Registering BlockManager BlockManagerId(driver, h12, 58572, None)
   	2022-10-12 08:42:26 INFO storage.BlockManagerMasterEndpoint: Registering block manager h12:58572 with 93.3 MiB RAM, BlockManagerId(driver, h12, 58572, None)
   	2022-10-12 08:42:26 INFO storage.BlockManagerMaster: Registered BlockManager BlockManagerId(driver, h12, 58572, None)
   	2022-10-12 08:42:26 INFO storage.BlockManager: Initialized BlockManager: BlockManagerId(driver, h12, 58572, None)
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@53f4c1e6{/metrics/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO history.SingleEventLogFileWriter: Logging events to hdfs://cluster1/cluster1/spark-logs/local-1665535346579.lz4.inprogress
   	2022-10-12 08:42:26 INFO internal.SharedState: loading hive config file: file:/opt/usdp-srv/srv/udp/2.0.0.0/spark/conf/hive-site.xml
   	2022-10-12 08:42:26 INFO internal.SharedState: spark.sql.warehouse.dir is not set, but hive.metastore.warehouse.dir is set. Setting spark.sql.warehouse.dir to the value of hive.metastore.warehouse.dir ('/cluster1/user/hive/warehouse/').
   	2022-10-12 08:42:26 INFO internal.SharedState: Warehouse path is '/cluster1/user/hive/warehouse/'.
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@135a8c6f{/SQL,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@280d4882{/SQL/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@56b859a6{/SQL/execution,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@42aae04d{/SQL/execution/json,null,AVAILABLE,@Spark}
   	2022-10-12 08:42:26 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@516462cc{/static/sql,null,AVAILABLE,@Spark}
   [INFO] 2022-10-12 00:42:28.065 +0000 -  -> 2022-10-12 08:42:27 INFO hive.HiveUtils: Initializing HiveMetastoreConnection version 2.3.7 using Spark classes.
   	2022-10-12 08:42:27 INFO client.HiveClientImpl: Warehouse location for Hive client (version 2.3.7) is /cluster1/user/hive/warehouse/
   	2022-10-12 08:42:27 INFO hive.metastore: Trying to connect to metastore with URI thrift://h10:9083
   	2022-10-12 08:42:27 INFO hive.metastore: Opened a connection to metastore, current connections: 1
   	2022-10-12 08:42:27 INFO hive.metastore: Connected to metastore.
   	Error in query: 
   	no viable alternative at input 'select \n'(line 2, pos 0)
   	
   	== SQL ==
   	select 
   	^^^
   	
   	2022-10-12 08:42:27 INFO server.AbstractConnector: Stopped Spark@64f1fd08{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
   	2022-10-12 08:42:27 INFO ui.SparkUI: Stopped Spark web UI at http://h12:4040
   [INFO] 2022-10-12 00:42:28.296 +0000 - process has exited, execute path:/tmp/dolphinscheduler/exec/process/7170590636544/7172773527040_35/121/137, processId:689910 ,exitStatusCode:1 ,processWaitForStatus:true ,processExitValue:1
   [INFO] 2022-10-12 00:42:29.066 +0000 -  -> 2022-10-12 08:42:28 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
   	2022-10-12 08:42:28 INFO memory.MemoryStore: MemoryStore cleared
   	2022-10-12 08:42:28 INFO storage.BlockManager: BlockManager stopped
   	2022-10-12 08:42:28 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
   	2022-10-12 08:42:28 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
   	2022-10-12 08:42:28 INFO spark.SparkContext: Successfully stopped SparkContext
   	2022-10-12 08:42:28 INFO util.ShutdownHookManager: Shutdown hook called
   	2022-10-12 08:42:28 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-69e6cf7a-7ef6-45ed-b2ff-3ce5ff5a3d89
   	2022-10-12 08:42:28 INFO util.ShutdownHookManager: Deleting directory /home/data/udp/2.0.0.0/spark/spark-c2a0103c-3876-445e-a12a-06798cb962da
   [INFO] 2022-10-12 00:42:29.067 +0000 - FINALIZE_SESSION
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@dolphinscheduler.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org