You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Beam JIRA Bot (Jira)" <ji...@apache.org> on 2021/01/09 17:13:05 UTC

[jira] [Updated] (BEAM-11224) Spark runner side inputs causing SIGNAL TERM

     [ https://issues.apache.org/jira/browse/BEAM-11224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Beam JIRA Bot updated BEAM-11224:
---------------------------------
    Labels: sideinput spark spark-runner stale-P2  (was: sideinput spark spark-runner)

> Spark runner side inputs causing SIGNAL TERM
> --------------------------------------------
>
>                 Key: BEAM-11224
>                 URL: https://issues.apache.org/jira/browse/BEAM-11224
>             Project: Beam
>          Issue Type: Bug
>          Components: runner-spark
>    Affects Versions: 2.25.0
>         Environment: Beam: 2.25
> SparkRunner: 2.25
> Java version: 11.0.9-ea
> Maven Compiler Source: 1.8
> Maven Compiler Target: 1.8
> Spark version: 2.4.7
>            Reporter: Thomas Li Fredriksen
>            Priority: P2
>              Labels: sideinput, spark, spark-runner, stale-P2
>          Time Spent: 10m
>  Remaining Estimate: 0h
>
> I wish to use side-inputs in order to pass some configuration to my pipeline, however the driver commands a shutdown after the `PCollectionView` has been created when running on my local spark-cluster (spark version 2.4.7, 1 master, 1 worker, running on localhost). This however works perfectly on the DirectRunner.
> I have attempted to strip the code to its bare essentials (see below). Still the issue persists when running on the spark cluster. DirectRunner still works fine.
> The spark-cluster does accept jobs, and I have sucessfully run a "hello-world" pipeline that completed without issue.
> What is happening here?
> Logs pasted below.
> {code}
>     // Pipeline
>     private static PipelineResult runPipeline(PipelineOptions options) {
>         Pipeline p = Pipeline.create(options);
>         PCollectionView<String> schema = p
>                 .apply("Dummy tabular schema builder", Create.of("This is a string"))
>                 .apply("Collect", View.asSingleton());
>         p
>                 .apply("Hello world", Create.of("Hello world"))
>                 .apply("Side input test", ParDo.of(DummyFn.builder().setSchemaView(schema).build()).withSideInput("schema", schema))
>                 .apply(ConsoleIO.create());
>         return p.run();
>     }
> {code}
> {code}
> // Simple FN that prints the side input
> @AutoValue
> public abstract class DummyFn extends DoFn<String, String> {
>     private final static Logger LOG = LoggerFactory.getLogger(DummyFn.class);
>     public static Builder builder() {
>         return new org.odp.beam.io.fn.AutoValue_DummyFn.Builder();
>     }
>     public abstract PCollectionView<String> getSchemaView();
>     @ProcessElement
>     public void processElement(@Element String element,
>                                OutputReceiver<String> out,
>                                ProcessContext context) {
>         String schema = context.sideInput(getSchemaView());
>         LOG.warn(schema.toString());
>         out.output(element.toUpperCase());
>     }
>     @AutoValue.Builder
>     public abstract static class Builder {
>         public abstract Builder setSchemaView(PCollectionView<String> value);
>         public abstract DummyFn build();
>     }
> }
> {code}
> {code}
> // Simple PTransform that prints the output of the toString-method
> public class ConsoleIO<T> extends PTransform<PCollection<T>, PDone> {
>     public static <T> ConsoleIO<T> create() {
>         return new ConsoleIO();
>     }
>     @Override
>     public PDone expand(PCollection<T> input) {
>         input
>                 .apply("Print elements", ParDo.of(new PrintElementFn<T>()));
>         return PDone.in(input.getPipeline());
>     }
>     static class PrintElementFn<T> extends DoFn<T, Void> {
>         @DoFn.ProcessElement
>         public void processElement(@Element T element, ProcessContext context) throws Exception {
>             System.out.println(element.toString());
>         }
>     }
> }
> {code}
> spark-submit output
> {code}
> $ spark-submit \
> --class org.odp.beam.extractors.CsvToCdfRawExtractor \
> --verbose \
> --driver-memory 4G \
> --executor-memory 4G \
> --total-executor-cores 4 \
> --deploy-mode client \
> --supervise \
> --conf spark.dynamicAllocation.enabled=false \
> --conf spark.network.timeout=420000 \
> --master spark://192.168.10.172:7077 \
> target/beam-poc-0.1-shaded.jar \
> --runner=SparkRunner
> Using properties file: null
> 20/11/10 15:46:44 WARN Utils: Your hostname, localhost.localdomain resolves to a loopback address: 127.0.0.1; using 192.168.10.172 instead (on interface enp7s0)
> 20/11/10 15:46:44 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
> WARNING: An illegal reflective access operation has occurred
> WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/jars/spark-unsafe_2.11-2.4.7.jar) to method java.nio.Bits.unaligned()
> WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform
> WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
> WARNING: All illegal access operations will be denied in a future release
> Parsed arguments:
>   master                  spark://192.168.10.172:7077
>   deployMode              client
>   executorMemory          4G
>   executorCores           null
>   totalExecutorCores      4
>   propertiesFile          null
>   driverMemory            4G
>   driverCores             null
>   driverExtraClassPath    null
>   driverExtraLibraryPath  null
>   driverExtraJavaOptions  null
>   supervise               true
>   queue                   null
>   numExecutors            null
>   files                   null
>   pyFiles                 null
>   archives                null
>   mainClass               org.odp.beam.extractors.CsvToCdfRawExtractor
>   primaryResource         file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar
>   name                    org.odp.beam.extractors.CsvToCdfRawExtractor
>   childArgs               [--runner=SparkRunner]
>   jars                    null
>   packages                null
>   packagesExclusions      null
>   repositories            null
>   verbose                 true
> Spark properties used, including those specified through
>  --conf and those from the properties file null:
>   (spark.network.timeout,420000)
>   (spark.driver.memory,4G)
>   (spark.dynamicAllocation.enabled,false)
>     
> 20/11/10 15:46:45 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
> Main class:
> org.odp.beam.extractors.CsvToCdfRawExtractor
> Arguments:
> --runner=SparkRunner
> Spark config:
> (spark.jars,file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar)
> (spark.app.name,org.odp.beam.extractors.CsvToCdfRawExtractor)
> (spark.cores.max,4)
> (spark.network.timeout,420000)
> (spark.driver.memory,4G)
> (spark.submit.deployMode,client)
> (spark.master,spark://192.168.10.172:7077)
> (spark.executor.memory,4G)
> (spark.dynamicAllocation.enabled,false)
> Classpath elements:
> file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar
> log4j:WARN No appenders could be found for logger (org.apache.beam.sdk.options.PipelineOptionsFactory).
> log4j:WARN Please initialize the log4j system properly.
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
> Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
> 20/11/10 15:46:46 INFO SparkContext: Running Spark version 2.4.7
> 20/11/10 15:46:47 INFO SparkContext: Submitted application: CsvToCdfRawExtractor
> 20/11/10 15:46:47 INFO SecurityManager: Changing view acls to: tom
> 20/11/10 15:46:47 INFO SecurityManager: Changing modify acls to: tom
> 20/11/10 15:46:47 INFO SecurityManager: Changing view acls groups to: 
> 20/11/10 15:46:47 INFO SecurityManager: Changing modify acls groups to: 
> 20/11/10 15:46:47 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(tom); groups with view permissions: Set(); users  with modify permissions: Set(tom); groups with modify permissions: Set()
> 20/11/10 15:46:47 INFO Utils: Successfully started service 'sparkDriver' on port 35103.
> 20/11/10 15:46:47 INFO SparkEnv: Registering MapOutputTracker
> 20/11/10 15:46:47 INFO SparkEnv: Registering BlockManagerMaster
> 20/11/10 15:46:47 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
> 20/11/10 15:46:47 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
> 20/11/10 15:46:47 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-58419068-d0ad-45c9-b90b-92b659dee1c3
> 20/11/10 15:46:47 INFO MemoryStore: MemoryStore started with capacity 2.2 GB
> 20/11/10 15:46:47 INFO SparkEnv: Registering OutputCommitCoordinator
> 20/11/10 15:46:47 INFO Utils: Successfully started service 'SparkUI' on port 4040.
> 20/11/10 15:46:47 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://fedora:4040
> 20/11/10 15:46:47 INFO SparkContext: Added JAR file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar at spark://fedora:35103/jars/beam-poc-0.1-shaded.jar with timestamp 1605019607514
> 20/11/10 15:46:47 INFO StandaloneAppClient$ClientEndpoint: Connecting to master spark://192.168.10.172:7077...
> 20/11/10 15:46:47 INFO TransportClientFactory: Successfully created connection to /192.168.10.172:7077 after 25 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:47 INFO StandaloneSchedulerBackend: Connected to Spark cluster with app ID app-20201110154647-0020
> 20/11/10 15:46:47 INFO StandaloneAppClient$ClientEndpoint: Executor added: app-20201110154647-0020/0 on worker-20201109144752-192.168.10.172-45535 (192.168.10.172:45535) with 4 core(s)
> 20/11/10 15:46:47 INFO StandaloneSchedulerBackend: Granted executor ID app-20201110154647-0020/0 on hostPort 192.168.10.172:45535 with 4 core(s), 4.0 GB RAM
> 20/11/10 15:46:47 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 33169.
> 20/11/10 15:46:47 INFO NettyBlockTransferService: Server created on fedora:33169
> 20/11/10 15:46:47 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
> 20/11/10 15:46:47 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20201110154647-0020/0 is now RUNNING
> 20/11/10 15:46:47 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO BlockManagerMasterEndpoint: Registering block manager fedora:33169 with 2.2 GB RAM, BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO StandaloneSchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.0
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Entering directly-translatable composite transform: 'Collect/Combine.GloballyAsSingletonView/Combine.globally(Singleton)'
> 20/11/10 15:46:48 INFO MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
> 20/11/10 15:46:48 INFO AggregatorsAccumulator: Instantiated aggregators accumulator: 
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Evaluating Read(CreateSource)
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Entering directly-translatable composite transform: 'Collect/Combine.GloballyAsSingletonView/Combine.globally(Singleton)'
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Evaluating Combine.globally(Singleton)
> 20/11/10 15:46:48 INFO SparkContext: Starting job: aggregate at GroupCombineFunctions.java:107
> 20/11/10 15:46:48 INFO DAGScheduler: Got job 0 (aggregate at GroupCombineFunctions.java:107) with 1 output partitions
> 20/11/10 15:46:48 INFO DAGScheduler: Final stage: ResultStage 0 (aggregate at GroupCombineFunctions.java:107)
> 20/11/10 15:46:48 INFO DAGScheduler: Parents of final stage: List()
> 20/11/10 15:46:48 INFO DAGScheduler: Missing parents: List()
> 20/11/10 15:46:48 INFO DAGScheduler: Submitting ResultStage 0 (Dummy tabular schema builder/Read(CreateSource).out Bounded[0] at RDD at SourceRDD.java:80), which has no missing parents
> 20/11/10 15:46:48 INFO MemoryStore: Block broadcast_0 stored as values in memory (estimated size 16.2 KB, free 2.2 GB)
> 20/11/10 15:46:48 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 6.8 KB, free 2.2 GB)
> 20/11/10 15:46:48 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on fedora:33169 (size: 6.8 KB, free: 2.2 GB)
> 20/11/10 15:46:48 INFO SparkContext: Created broadcast 0 from broadcast at DAGScheduler.scala:1184
> 20/11/10 15:46:48 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (Dummy tabular schema builder/Read(CreateSource).out Bounded[0] at RDD at SourceRDD.java:80) (first 15 tasks are for partitions Vector(0))
> 20/11/10 15:46:48 INFO TaskSchedulerImpl: Adding task set 0.0 with 1 tasks
> 20/11/10 15:46:49 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Registered executor NettyRpcEndpointRef(spark-client://Executor) (192.168.10.172:48382) with ID 0
> 20/11/10 15:46:49 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, 192.168.10.172, executor 0, partition 0, PROCESS_LOCAL, 8546 bytes)
> 20/11/10 15:46:49 INFO BlockManagerMasterEndpoint: Registering block manager 192.168.10.172:43781 with 2.2 GB RAM, BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:50 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 192.168.10.172:43781 (size: 6.8 KB, free: 2.2 GB)
> 20/11/10 15:46:51 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 2056 ms on 192.168.10.172 (executor 0) (1/1)
> 20/11/10 15:46:51 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool 
> 20/11/10 15:46:51 INFO DAGScheduler: ResultStage 0 (aggregate at GroupCombineFunctions.java:107) finished in 3.091 s
> 20/11/10 15:46:51 INFO DAGScheduler: Job 0 finished: aggregate at GroupCombineFunctions.java:107, took 3.132405 s
> 20/11/10 15:46:51 INFO SparkRunner$Evaluator: Evaluating org.apache.beam.sdk.transforms.View$VoidKeyToMultimapMaterialization$VoidKeyToMultimapMaterializationDoFn@14924f41
> 20/11/10 15:46:51 INFO SparkRunner$Evaluator: Evaluating View.CreatePCollectionView
> 20/11/10 15:46:51 INFO SparkContext: Invoking stop() from shutdown hook
> 20/11/10 15:46:51 INFO SparkUI: Stopped Spark web UI at http://fedora:4040
> 20/11/10 15:46:51 INFO StandaloneSchedulerBackend: Shutting down all executors
> 20/11/10 15:46:51 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asking each executor to shut down
> 20/11/10 15:46:51 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
> 20/11/10 15:46:51 INFO MemoryStore: MemoryStore cleared
> 20/11/10 15:46:51 INFO BlockManager: BlockManager stopped
> 20/11/10 15:46:51 INFO BlockManagerMaster: BlockManagerMaster stopped
> 20/11/10 15:46:51 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
> 20/11/10 15:46:51 INFO SparkContext: Successfully stopped SparkContext
> 20/11/10 15:46:51 INFO ShutdownHookManager: Shutdown hook called
> 20/11/10 15:46:51 INFO ShutdownHookManager: Deleting directory /tmp/spark-665a903f-22db-497e-989f-a5ca3e0635e2
> 20/11/10 15:46:51 INFO ShutdownHookManager: Deleting directory /tmp/spark-d4b5a04f-f6a3-48ff-b229-4eb966151d86
> {code}
> stderr from spark worker:
> {code}
> Spark Executor Command: "/usr/lib/jvm/java-11-openjdk-11.0.9.6-0.0.ea.fc33.x86_64/bin/java" "-cp" "/home/tom/app/spark/spark/conf/:/home/tom/app/spark/spark/jars/*" "-Xmx4096M" "-Dspark.driver.port=35103" "-Dspark.network.timeout=420000" "org.apache.spark.executor.CoarseGrainedExecutorBackend" "--driver-url" "spark://CoarseGrainedScheduler@fedora:35103" "--executor-id" "0" "--hostname" "192.168.10.172" "--cores" "4" "--app-id" "app-20201110154647-0020" "--worker-url" "spark://Worker@192.168.10.172:45535"
> ========================================
> Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
> 20/11/10 15:46:48 INFO CoarseGrainedExecutorBackend: Started daemon with process name: 360353@localhost.localdomain
> 20/11/10 15:46:48 INFO SignalUtils: Registered signal handler for TERM
> 20/11/10 15:46:48 INFO SignalUtils: Registered signal handler for HUP
> 20/11/10 15:46:48 INFO SignalUtils: Registered signal handler for INT
> 20/11/10 15:46:48 WARN Utils: Your hostname, localhost.localdomain resolves to a loopback address: 127.0.0.1; using 192.168.10.172 instead (on interface enp7s0)
> 20/11/10 15:46:48 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
> WARNING: An illegal reflective access operation has occurred
> WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/jars/spark-unsafe_2.11-2.4.7.jar) to method java.nio.Bits.unaligned()
> WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform
> WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
> WARNING: All illegal access operations will be denied in a future release
> 20/11/10 15:46:48 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
> 20/11/10 15:46:48 INFO SecurityManager: Changing view acls to: tom
> 20/11/10 15:46:48 INFO SecurityManager: Changing modify acls to: tom
> 20/11/10 15:46:48 INFO SecurityManager: Changing view acls groups to: 
> 20/11/10 15:46:48 INFO SecurityManager: Changing modify acls groups to: 
> 20/11/10 15:46:48 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(tom); groups with view permissions: Set(); users  with modify permissions: Set(tom); groups with modify permissions: Set()
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created connection to fedora/192.168.10.172:35103 after 54 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:49 INFO SecurityManager: Changing view acls to: tom
> 20/11/10 15:46:49 INFO SecurityManager: Changing modify acls to: tom
> 20/11/10 15:46:49 INFO SecurityManager: Changing view acls groups to: 
> 20/11/10 15:46:49 INFO SecurityManager: Changing modify acls groups to: 
> 20/11/10 15:46:49 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(tom); groups with view permissions: Set(); users  with modify permissions: Set(tom); groups with modify permissions: Set()
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created connection to fedora/192.168.10.172:35103 after 4 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:49 INFO DiskBlockManager: Created local directory at /tmp/spark-0e47fa97-8714-4e8e-950e-b1032fe36995/executor-e7667d04-198d-4144-8897-ddada0bfd1de/blockmgr-019262b3-4d3e-4158-b984-ff85c0846191
> 20/11/10 15:46:49 INFO MemoryStore: MemoryStore started with capacity 2.2 GB
> 20/11/10 15:46:49 INFO CoarseGrainedExecutorBackend: Connecting to driver: spark://CoarseGrainedScheduler@fedora:35103
> 20/11/10 15:46:49 INFO WorkerWatcher: Connecting to worker spark://Worker@192.168.10.172:45535
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created connection to /192.168.10.172:45535 after 2 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:49 INFO WorkerWatcher: Successfully connected to spark://Worker@192.168.10.172:45535
> 20/11/10 15:46:49 INFO CoarseGrainedExecutorBackend: Successfully registered with driver
> 20/11/10 15:46:49 INFO Executor: Starting executor ID 0 on host 192.168.10.172
> 20/11/10 15:46:49 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43781.
> 20/11/10 15:46:49 INFO NettyBlockTransferService: Server created on 192.168.10.172:43781
> 20/11/10 15:46:49 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
> 20/11/10 15:46:49 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:49 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:49 INFO BlockManager: Initialized BlockManager: BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:49 INFO CoarseGrainedExecutorBackend: Got assigned task 0
> 20/11/10 15:46:49 INFO Executor: Running task 0.0 in stage 0.0 (TID 0)
> 20/11/10 15:46:49 INFO Executor: Fetching spark://fedora:35103/jars/beam-poc-0.1-shaded.jar with timestamp 1605019607514
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created connection to fedora/192.168.10.172:35103 after 2 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:49 INFO Utils: Fetching spark://fedora:35103/jars/beam-poc-0.1-shaded.jar to /tmp/spark-0e47fa97-8714-4e8e-950e-b1032fe36995/executor-e7667d04-198d-4144-8897-ddada0bfd1de/spark-62556d02-a044-4c2c-8f97-c7f25ef3e337/fetchFileTemp6325880319900581024.tmp
> 20/11/10 15:46:49 INFO Utils: Copying /tmp/spark-0e47fa97-8714-4e8e-950e-b1032fe36995/executor-e7667d04-198d-4144-8897-ddada0bfd1de/spark-62556d02-a044-4c2c-8f97-c7f25ef3e337/2058038551605019607514_cache to /home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/work/app-20201110154647-0020/0/./beam-poc-0.1-shaded.jar
> 20/11/10 15:46:50 INFO Executor: Adding file:/home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/work/app-20201110154647-0020/0/./beam-poc-0.1-shaded.jar to class loader
> 20/11/10 15:46:50 INFO TorrentBroadcast: Started reading broadcast variable 0
> 20/11/10 15:46:50 INFO TransportClientFactory: Successfully created connection to fedora/192.168.10.172:33169 after 2 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:50 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 6.8 KB, free 2.2 GB)
> 20/11/10 15:46:50 INFO TorrentBroadcast: Reading broadcast variable 0 took 112 ms
> 20/11/10 15:46:50 INFO MemoryStore: Block broadcast_0 stored as values in memory (estimated size 16.2 KB, free 2.2 GB)
> 20/11/10 15:46:51 INFO Executor: Finished task 0.0 in stage 0.0 (TID 0). 6312 bytes result sent to driver
> 20/11/10 15:46:51 INFO CoarseGrainedExecutorBackend: Driver commanded a shutdown
> 20/11/10 15:46:51 ERROR CoarseGrainedExecutorBackend: RECEIVED SIGNAL TERM
> tdown
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)