[2022-11-14T11:06:34,258] [INFO ] o.t.d.s.FullRunJob$ - [ option: runtimeFlow runtimeFlow [ARG] :: Base64 encoded runtimeflow ] [2022-11-14T11:06:34,264] [INFO ] o.t.d.s.FullRunJob$ - [ option: pluginIds pluginIds [ARG] :: Ignored ] [2022-11-14T11:06:34,264] [INFO ] o.t.d.s.FullRunJob$ - [ option: metricsSinkUrl metricsSinkUrl [ARG] :: Metrics sink URL ] [2022-11-14T11:06:34,264] [INFO ] o.t.d.s.FullRunJob$ - [ option: metricsRefreshInterval metricsRefreshInterval [ARG] :: Metrics refresh interval ] [2022-11-14T11:06:34,264] [INFO ] o.t.d.s.FullRunJob$ - Reading RuntimeFlow from runtimeFlow argument [2022-11-14T11:06:34,379] [INFO ] v.JsonUtils$ - Getting Vault key name from the received payload [2022-11-14T11:06:34,435] [INFO ] v.VaultProxy - Client doesn't provide any 'sensitiveFields', it will apply default regex mechanism to decrypt. [2022-11-14T11:06:34,439] [INFO ] v.VaultProxy - The payload doesn't contain any value to be encrypted [2022-11-14T11:06:35,157] [INFO ] o.t.d.s.u.MetricsHelper$ - Using Metrics parameters from Livy environment variables [2022-11-14T11:06:35,230] [INFO ] o.t.d.s.u.ReplicateContentHelper$ - Disabling Beam splittable DoFns for compatibility reasons [2022-11-14T11:06:35,240] [INFO ] o.t.d.s.FullRunJob$ - Done parsing and validating the RuntimeFlow, starting the job... [2022-11-14T11:06:35,411] [INFO ] o.t.s.c.c.ContainerManager - No TALEND-INF/plugins.properties found, will use file resolution [2022-11-14T11:06:35,469] [INFO ] o.t.s.c.r.m.ComponentManager - ComponentManager version: 1.51.0-SNAPSHOT [2022-11-14T11:06:35,473] [INFO ] o.t.s.c.r.m.ComponentManager - Creating the contextual ComponentManager instance (classloader=jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27, jvm=19385@793948bf7ece) [2022-11-14T11:06:35,475] [INFO ] o.t.s.c.r.m.ComponentManager - Components: [] [2022-11-14T11:06:35,475] [INFO ] o.t.s.c.r.m.ComponentManager - Created the contextual ComponentManager instance (classloader=jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27, jvm=19385@793948bf7ece) [2022-11-14T11:06:35,479] [INFO ] o.t.d.t.ConnectorsRepoContainerFinder - Initializing from /opt/talend/connectors [2022-11-14T11:06:35,488] [INFO ] o.t.s.c.c.ContainerManager - Creating module /opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar (from /opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar, location=/opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar) [2022-11-14T11:06:35,765] [INFO ] o.t.s.c.r.m.ComponentManager - Added @Service class org.talend.components.localio.fixed.FixedFlowInputService for container-id=localio [2022-11-14T11:06:35,809] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.localio.fixed.FixedFlowInputRuntime for container-id=localio [2022-11-14T11:06:35,820] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.localio.devnull.DevNullOutputRuntime for container-id=localio [2022-11-14T11:06:35,830] [INFO ] o.t.c.e.r.i.RegisteredComponentExtension - RegisteredComponentExtension extension will load CustomComponentExtension... [2022-11-14T11:06:35,833] [INFO ] o.t.s.c.c.ContainerManager - Created container localio [2022-11-14T11:06:35,833] [INFO ] o.t.s.c.r.m.ComponentManager - Adding plugin: /opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar, as localio [2022-11-14T11:06:35,834] [INFO ] o.t.s.c.c.ContainerManager - Creating module /opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar (from /opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar, location=/opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar) [2022-11-14T11:06:35,974] [INFO ] o.t.s.c.r.m.ComponentManager - Added @Service class org.talend.components.processing.ProcessingService for container-id=processing-beam [2022-11-14T11:06:36,025] [INFO ] o.t.s.c.r.m.ComponentManager - Added @Service class org.talend.components.processing.join.JoinService for container-id=processing-beam [2022-11-14T11:06:36,040] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.aggregate.Aggregate for container-id=processing-beam [2022-11-14T11:06:36,045] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.applyschema.ApplySchema for container-id=processing-beam [2022-11-14T11:06:36,051] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.fieldremover.FieldRemover for container-id=processing-beam [2022-11-14T11:06:36,079] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.fieldselector.FieldSelector for container-id=processing-beam [2022-11-14T11:06:36,097] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.filter.Filter for container-id=processing-beam [2022-11-14T11:06:36,101] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.filter.SemanticFilter for container-id=processing-beam [2022-11-14T11:06:36,104] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.filter.TechnicalFilter for container-id=processing-beam [2022-11-14T11:06:36,109] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.flatten.Flatten for container-id=processing-beam [2022-11-14T11:06:36,122] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.join.JoinJobBuilder for container-id=processing-beam [2022-11-14T11:06:36,127] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.limit.Limit for container-id=processing-beam [2022-11-14T11:06:36,128] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.limit.Sample for container-id=processing-beam [2022-11-14T11:06:36,135] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.python.Python for container-id=processing-beam [2022-11-14T11:06:36,139] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.replicate.Replicate for container-id=processing-beam [2022-11-14T11:06:36,153] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.sampling.Sampling for container-id=processing-beam [2022-11-14T11:06:36,158] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.typeconverter.TypeConverter for container-id=processing-beam [2022-11-14T11:06:36,167] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.window.Window for container-id=processing-beam [2022-11-14T11:06:36,180] [INFO ] o.t.c.e.r.i.RegisteredComponentExtension - RegisteredComponentExtension extension will load CustomComponentExtension... [2022-11-14T11:06:36,181] [INFO ] o.t.s.c.c.ContainerManager - Created container processing-beam [2022-11-14T11:06:36,181] [INFO ] o.t.s.c.r.m.ComponentManager - Adding plugin: /opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar, as processing-beam [2022-11-14T11:06:37,656] [INFO ] o.t.d.b.c.r.RuntimeFlowBeamCompilerContext - Creating Beam Pipeline by: class com.sun.proxy.$Proxy104 [2022-11-14T11:06:37,815] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:38,066] [INFO ] o.t.d.b.c.BeamCompiler - Start to resolve runtime for component bG9jYWxpbyNMb2NhbElPI0ZpeGVkRmxvd0lucHV0UnVudGltZQ (id=hvu7itwhwixpbzqwmngcmfhz) [2022-11-14T11:06:38,072] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - ====== Get runtime of a component ====== [2022-11-14T11:06:38,077] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - localio#LocalIO#FixedFlowInputRuntime [2022-11-14T11:06:38,079] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Selected a component of type: MAPPER [2022-11-14T11:06:38,084] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Component version is: 1 [2022-11-14T11:06:38,316] [INFO ] o.t.d.b.c.BeamCompiler - Tacokit component runtime: MapperTransform [2022-11-14T11:06:38,375] [WARN ] o.t.s.c.r.s.EnhancedObjectInputStream$Defaults - talend.component.runtime.serialization.java.inputstream.whitelist system property not set, will use default blacklist but this is not considered as a secure setup. Blacklisted packages: [com.sun.rowset., org.apache.ibatis., org.apache.xalan, com.sun.org.apache.xalan.internal.xsltc.trax., java.rmi.server., oracle.jdbc., java.lang.Process, org.springframework.beans.factory.config., org.slf4j.ext., java.util.logging., com.sun.org.apache.bcel.internal.util., com.sun.deploy.security.ruleset., org.apache.axis2.jaxws.spi.handler., org.apache.commons.collections4.functors., org.apache.commons.collections.functors., org.hibernate.jmx., flex.messaging.util.concurrent., org.apache.openjpa.ee., org.codehaus.groovy.runtime., org.apache.axis2.transport.jms., org.apache.tomcat.dbcp.dbcp2., jodd.db.connection., org.jboss.util.propertyeditor.] [2022-11-14T11:06:38,778] [INFO ] o.t.d.b.c.BeamCompiler - Start to resolve runtime for component cHJvY2Vzc2luZy1iZWFtI1Byb2Nlc3NpbmcjUHl0aG9u (id=hkfejapx6goigmclkausdf2m) [2022-11-14T11:06:38,783] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - ====== Get runtime of a component ====== [2022-11-14T11:06:38,783] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - processing-beam#Processing#Python [2022-11-14T11:06:38,783] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Selected a component of type: PROCESSOR [2022-11-14T11:06:38,787] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Component version is: 1 [2022-11-14T11:06:38,800] [WARN ] o.t.s.c.r.b.s.BeamComponentExtension - Creating a 'interface org.talend.sdk.component.runtime.output.Processor' instance for 'Processing#Python', this must be unwrapped before being used [2022-11-14T11:06:38,806] [WARN ] o.t.s.c.r.b.s.BeamComponentExtension - Creating a 'interface org.talend.sdk.component.runtime.output.Processor' instance for 'Processing#Python', this must be unwrapped before being used [2022-11-14T11:06:38,807] [INFO ] o.t.d.b.c.BeamCompiler - Tacokit component runtime: Python [2022-11-14T11:06:38,848] [INFO ] o.t.d.b.c.BeamCompiler - Start to resolve runtime for component bG9jYWxpbyNMb2NhbElPI0Rldk51bGxPdXRwdXRSdW50aW1l (id=xavz5scolwhovwcxpkntvnp1) [2022-11-14T11:06:38,850] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - ====== Get runtime of a component ====== [2022-11-14T11:06:38,850] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - localio#LocalIO#DevNullOutputRuntime [2022-11-14T11:06:38,850] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Selected a component of type: PROCESSOR [2022-11-14T11:06:38,851] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Component version is: 1 [2022-11-14T11:06:38,873] [INFO ] o.t.d.b.c.BeamCompiler - Tacokit component runtime: org.talend.sdk.component.runtime.output.ProcessorImpl@bcd4fffe [2022-11-14T11:06:38,913] [WARN ] o.t.d.b.c.r.RuntimeFlowBeamCompilerContext - Not able to parse 'metrics.socketTimeoutMs', default is 15000 [2022-11-14T11:06:38,913] [WARN ] o.t.d.b.c.r.RuntimeFlowBeamCompilerContext - Not able to parse 'metrics.connectionTimeoutMs', default is 15000 [2022-11-14T11:06:39,145] [INFO ] o.a.s.SparkContext - Running Spark version 3.2.0 [2022-11-14T11:06:39,321] [WARN ] o.a.h.u.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable [2022-11-14T11:06:39,424] [INFO ] o.a.s.r.ResourceUtils - ============================================================== [2022-11-14T11:06:39,424] [INFO ] o.a.s.r.ResourceUtils - No custom resources configured for spark.driver. [2022-11-14T11:06:39,425] [INFO ] o.a.s.r.ResourceUtils - ============================================================== [2022-11-14T11:06:39,425] [INFO ] o.a.s.SparkContext - Submitted application: pipeline for proc-python2-reading-file processor [2022-11-14T11:06:39,446] [INFO ] o.a.s.r.ResourceProfile - Default ResourceProfile created, executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , memory -> name: memory, amount: 1024, script: , vendor: , offHeap -> name: offHeap, amount: 0, script: , vendor: ), task resources: Map(cpus -> name: cpus, amount: 1.0) [2022-11-14T11:06:39,451] [INFO ] o.a.s.r.ResourceProfile - Limiting resource is cpu [2022-11-14T11:06:39,452] [INFO ] o.a.s.r.ResourceProfileManager - Added ResourceProfile id: 0 [2022-11-14T11:06:39,507] [INFO ] o.a.s.SecurityManager - Changing view acls to: root [2022-11-14T11:06:39,508] [INFO ] o.a.s.SecurityManager - Changing modify acls to: root [2022-11-14T11:06:39,508] [INFO ] o.a.s.SecurityManager - Changing view acls groups to: [2022-11-14T11:06:39,509] [INFO ] o.a.s.SecurityManager - Changing modify acls groups to: [2022-11-14T11:06:39,509] [INFO ] o.a.s.SecurityManager - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); groups with view permissions: Set(); users with modify permissions: Set(root); groups with modify permissions: Set() [2022-11-14T11:06:39,842] [INFO ] o.a.s.SparkEnv - Registering MapOutputTracker [2022-11-14T11:06:39,873] [INFO ] o.a.s.SparkEnv - Registering BlockManagerMaster [2022-11-14T11:06:39,896] [INFO ] o.a.s.SparkEnv - Registering BlockManagerMasterHeartbeat [2022-11-14T11:06:39,969] [INFO ] o.a.s.SparkEnv - Registering OutputCommitCoordinator [2022-11-14T11:06:40,064] [INFO ] o.a.s.e.Executor - Starting executor ID driver on host 793948bf7ece [2022-11-14T11:06:40,148] [INFO ] o.s.j.u.log - Logging initialized @7331ms to org.sparkproject.jetty.util.log.Slf4jLog [2022-11-14T11:06:40,205] [INFO ] o.t.d.b.c.BeamJob - Context variables initialization for this execution: [2022-11-14T11:06:40,223] [INFO ] o.t.c.c.d.p.DictionaryProvider - Loading registry from /tmp/dict/12e4d32a-69ea-4956-b382-dbb5704da353/indexFolder [2022-11-14T11:06:40,225] [INFO ] o.t.c.c.d.p.DictionaryProvider - The dictionary '0' can't be found in the local registry - creating it [2022-11-14T11:06:40,259] [INFO ] o.t.d.s.a.CategoryRegistryManager - Loading categories from local registry. [2022-11-14T11:06:40,434] [INFO ] o.t.d.s.i.ClassPathDirectory - Opening 'file:///tmp/dict/indexFolder/shared/prod/dictionary/' ... [2022-11-14T11:06:40,434] [INFO ] o.t.d.s.i.ClassPathDirectory - Opening 'file:///tmp/dict/indexFolder/shared/prod/keyword/' ... [2022-11-14T11:06:40,739] [INFO ] o.t.d.s.FullRunJob$ - Got dictionary 0 from MavenDictionaryProvider. [2022-11-14T11:06:40,739] [WARN ] o.t.d.s.FullRunJob$ - DQ rules is disabled for now, pending TPRUN-4490 [2022-11-14T11:06:40,743] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:40,748] [WARN ] o.a.b.s.Pipeline - The following transforms do not have stable unique names: , ParDo(VolumeOnLink), ParDo(NumberRecordsOnLink) [2022-11-14T11:06:40,748] [INFO ] o.a.b.r.s.SparkRunner - Executing pipeline using the SparkRunner. [2022-11-14T11:06:41,032] [INFO ] o.a.b.r.s.t.SparkContextFactory - Using a provided Spark Context [2022-11-14T11:06:41,256] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:41,295] [INFO ] o.a.b.r.s.m.MetricsAccumulator - Instantiated metrics accumulator: MetricQueryResults() [2022-11-14T11:06:41,298] [INFO ] o.a.b.r.s.a.AggregatorsAccumulator - Instantiated aggregators accumulator: [2022-11-14T11:06:41,302] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating Read(BoundedSourceImpl) [2022-11-14T11:06:41,378] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.util.builder.tacokit.transform.RecordToIndexedRecord@31773d5b [2022-11-14T11:06:41,383] [INFO ] o.t.d.b.c.BeamJob - Metrics are enabled [2022-11-14T11:06:41,535] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$NumberRecordsOnLink@71dac63f [2022-11-14T11:06:41,549] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$VolumeOnLink@5db6083b [2022-11-14T11:06:41,563] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.components.processing.python.PythonDoFn@c11ec9b [2022-11-14T11:06:41,578] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$NumberRecordsOnLink@118cbb26 [2022-11-14T11:06:41,590] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$VolumeOnLink@36a59d8 [2022-11-14T11:06:41,603] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.util.builder.tacokit.transform.IndexedRecordToRecord@3299e315 [2022-11-14T11:06:41,615] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.sdk.component.runtime.beam.transform.RecordNormalizer@10007fdc [2022-11-14T11:06:41,628] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.sdk.component.runtime.beam.TalendIO$WriteFn@699fe19f [2022-11-14T11:06:41,689] [INFO ] o.a.s.SparkContext - Starting job: foreach at BoundedDataset.java:127 [2022-11-14T11:06:41,709] [INFO ] o.a.s.s.DAGScheduler - Got job 0 (foreach at BoundedDataset.java:127) with 5 output partitions [2022-11-14T11:06:41,711] [INFO ] o.a.s.s.DAGScheduler - Final stage: ResultStage 0 (foreach at BoundedDataset.java:127) [2022-11-14T11:06:41,711] [INFO ] o.a.s.s.DAGScheduler - Parents of final stage: List() [2022-11-14T11:06:41,715] [INFO ] o.a.s.s.DAGScheduler - Missing parents: List() [2022-11-14T11:06:41,723] [INFO ] o.a.s.s.DAGScheduler - Submitting ResultStage 0 (3/Processor_LocalIO#DevNullOutputRuntime/ParDo(Write)/ParMultiDo(Write).output MapPartitionsRDD[27] at values at TransformTranslator.java:442), which has no missing parents [2022-11-14T11:06:41,962] [INFO ] o.a.s.SparkContext - Created broadcast 0 from broadcast at DAGScheduler.scala:1427 [2022-11-14T11:06:41,975] [INFO ] o.a.s.s.DAGScheduler - Submitting 5 missing tasks from ResultStage 0 (3/Processor_LocalIO#DevNullOutputRuntime/ParDo(Write)/ParMultiDo(Write).output MapPartitionsRDD[27] at values at TransformTranslator.java:442) (first 15 tasks are for partitions Vector(0, 1, 2, 3, 4)) [2022-11-14T11:06:41,976] [INFO ] o.a.s.s.TaskSchedulerImpl - Adding task set 0.0 with 5 tasks resource profile 0 [2022-11-14T11:06:42,031] [INFO ] o.a.s.s.TaskSetManager - Starting task 0.0 in stage 0.0 (TID 0) (793948bf7ece, executor driver, partition 0, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:06:42,034] [INFO ] o.a.s.s.TaskSetManager - Starting task 1.0 in stage 0.0 (TID 1) (793948bf7ece, executor driver, partition 1, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:06:42,039] [INFO ] o.a.s.s.TaskSetManager - Starting task 2.0 in stage 0.0 (TID 2) (793948bf7ece, executor driver, partition 2, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:06:42,042] [INFO ] o.a.s.s.TaskSetManager - Starting task 3.0 in stage 0.0 (TID 3) (793948bf7ece, executor driver, partition 3, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:06:42,056] [INFO ] o.a.s.e.Executor - Running task 0.0 in stage 0.0 (TID 0) [2022-11-14T11:06:42,057] [INFO ] o.a.s.e.Executor - Running task 2.0 in stage 0.0 (TID 2) [2022-11-14T11:06:42,058] [INFO ] o.a.s.e.Executor - Running task 1.0 in stage 0.0 (TID 1) [2022-11-14T11:06:42,063] [INFO ] o.a.s.e.Executor - Running task 3.0 in stage 0.0 (TID 3) [2022-11-14T11:06:42,150] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:42,150] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:42,153] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:42,156] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:43,023] [INFO ] o.a.s.e.Executor - Finished task 0.0 in stage 0.0 (TID 0). 7981 bytes result sent to driver [2022-11-14T11:06:43,024] [INFO ] o.a.s.e.Executor - Finished task 1.0 in stage 0.0 (TID 1). 7981 bytes result sent to driver [2022-11-14T11:06:43,025] [INFO ] o.a.s.e.Executor - Finished task 3.0 in stage 0.0 (TID 3). 8024 bytes result sent to driver [2022-11-14T11:06:43,027] [INFO ] o.a.s.s.TaskSetManager - Starting task 4.0 in stage 0.0 (TID 4) (793948bf7ece, executor driver, partition 4, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:06:43,027] [INFO ] o.a.s.e.Executor - Finished task 2.0 in stage 0.0 (TID 2). 7981 bytes result sent to driver [2022-11-14T11:06:43,037] [INFO ] o.a.s.e.Executor - Running task 4.0 in stage 0.0 (TID 4) [2022-11-14T11:06:43,038] [INFO ] o.a.s.s.TaskSetManager - Finished task 0.0 in stage 0.0 (TID 0) in 1018 ms on 793948bf7ece (executor driver) (1/5) [2022-11-14T11:06:43,061] [INFO ] o.a.s.s.TaskSetManager - Finished task 3.0 in stage 0.0 (TID 3) in 1019 ms on 793948bf7ece (executor driver) (2/5) [2022-11-14T11:06:43,061] [INFO ] o.a.s.s.TaskSetManager - Finished task 1.0 in stage 0.0 (TID 1) in 1027 ms on 793948bf7ece (executor driver) (3/5) [2022-11-14T11:06:43,066] [INFO ] o.a.s.s.TaskSetManager - Finished task 2.0 in stage 0.0 (TID 2) in 1027 ms on 793948bf7ece (executor driver) (4/5) [2022-11-14T11:06:43,070] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:06:48,851] [ERROR] o.a.s.e.Executor - Exception in task 4.0 in stage 0.0 (TID 4) org.apache.beam.sdk.util.UserCodeException: org.talend.components.processing.ProcessingErrorCode$TalendMsgRuntimeException: Access denied. at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.talend.components.processing.python.PythonDoFn$DoFnInvoker.invokeProcessElement(Unknown Source) ~[?:?] at org.apache.beam.runners.core.SimpleDoFnRunner.invokeProcessElement(SimpleDoFnRunner.java:228) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.apache.beam.runners.core.SimpleDoFnRunner.processElement(SimpleDoFnRunner.java:187) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.DoFnRunnerWithMetrics.processElement(DoFnRunnerWithMetrics.java:65) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.SparkProcessContext$ProcCtxtIterator.computeNext(SparkProcessContext.java:140) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:141) ~[beam-vendor-guava-26_0-jre-0.1.jar:?] at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:136) ~[beam-vendor-guava-26_0-jre-0.1.jar:?] at scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:45) ~[scala-library-2.12.15.jar:?] at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:513) ~[scala-library-2.12.15.jar:?] at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460) ~[scala-library-2.12.15.jar:?] at scala.collection.convert.Wrappers$IteratorWrapper.hasNext(Wrappers.scala:32) ~[scala-library-2.12.15.jar:?] at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:128) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:63) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsToPair$1(JavaRDDLike.scala:186) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$mapPartitions$2(RDD.scala:863) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$mapPartitions$2$adapted(RDD.scala:863) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.Task.run(Task.scala:131) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) [spark-core_2.12-3.2.0.jar:3.2.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] Caused by: org.talend.components.processing.ProcessingErrorCode$TalendMsgRuntimeException: Access denied. at org.talend.components.processing.ProcessingErrorCode.createAccessDeniedException(ProcessingErrorCode.java:147) ~[?:?] at org.talend.components.processing.python.PythonDoFn.map(PythonDoFn.java:197) ~[?:?] at org.talend.components.processing.python.PythonDoFn.processElement(PythonDoFn.java:78) ~[?:?] [2022-11-14T11:06:48,901] [WARN ] o.a.s.ThrowableSerializationWrapper - Task exception could not be deserialized java.lang.ClassNotFoundException: org.talend.components.processing.ProcessingErrorCode$TalendMsgRuntimeException at jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581) ~[?:?] at jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) ~[?:?] at java.lang.ClassLoader.loadClass(ClassLoader.java:522) ~[?:?] at java.lang.Class.forName0(Native Method) ~[?:?] at java.lang.Class.forName(Class.java:398) ~[?:?] at org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:68) ~[spark-core_2.12-3.2.0.jar:3.2.0] at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1986) ~[?:?] at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1853) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2184) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2479) ~[?:?] at java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:624) ~[?:?] at java.lang.Throwable.readObject(Throwable.java:896) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1046) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2340) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:488) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:446) ~[?:?] at org.apache.spark.ThrowableSerializationWrapper.readObject(TaskEndReason.scala:202) ~[spark-core_2.12-3.2.0.jar:3.2.0] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1046) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2340) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2479) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2373) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2479) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2373) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:488) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:446) ~[?:?] at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:76) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:115) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.TaskResultGetter.$anonfun$enqueueFailedTask$2(TaskResultGetter.scala:141) ~[spark-core_2.12-3.2.0.jar:3.2.0] at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) ~[scala-library-2.12.15.jar:?] at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.TaskResultGetter.$anonfun$enqueueFailedTask$1(TaskResultGetter.scala:137) ~[spark-core_2.12-3.2.0.jar:3.2.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] [2022-11-14T11:06:48,903] [ERROR] o.a.s.s.TaskResultGetter - Could not deserialize TaskEndReason: ClassNotFound with classloader jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27 [2022-11-14T11:06:48,905] [WARN ] o.a.s.s.TaskSetManager - Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason [2022-11-14T11:06:48,906] [ERROR] o.a.s.s.TaskSetManager - Task 4 in stage 0.0 failed 1 times; aborting job [2022-11-14T11:06:48,907] [INFO ] o.a.s.s.TaskSchedulerImpl - Removed TaskSet 0.0, whose tasks have all completed, from pool [2022-11-14T11:06:48,909] [INFO ] o.a.s.s.TaskSchedulerImpl - Cancelling stage 0 [2022-11-14T11:06:48,909] [INFO ] o.a.s.s.TaskSchedulerImpl - Killing all running tasks in stage 0: Stage cancelled [2022-11-14T11:06:48,911] [INFO ] o.a.s.s.DAGScheduler - ResultStage 0 (foreach at BoundedDataset.java:127) failed in 7.160 s due to Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: [2022-11-14T11:06:48,915] [INFO ] o.a.s.s.DAGScheduler - Job 0 failed: foreach at BoundedDataset.java:127, took 7.226630 s [2022-11-14T11:06:48,917] [INFO ] o.t.d.b.c.r.ServerSocketAvroSchemaRegistry - Shutdown registry containing {} [2022-11-14T11:06:48,918] [ERROR] o.t.d.s.FullRunJob$ - DATA_STREAMS_ERROR_MSG:Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: [2022-11-14T11:06:48,918] [ERROR] o.t.d.s.FullRunJob$ - Original Throwable: java.lang.RuntimeException: org.apache.spark.SparkException: Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: at org.apache.beam.runners.spark.SparkPipelineResult.runtimeExceptionFrom(SparkPipelineResult.java:60) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkPipelineResult.beamExceptionFrom(SparkPipelineResult.java:77) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:104) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:92) ~[beam-runners-spark-3-2.36.0.jar:?] at org.talend.datastreams.beam.compiler.BeamJob.waitUntilFinish(BeamJob.java:211) ~[data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.beam.compiler.BeamJob.fullRun(BeamJob.java:180) ~[data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob$.runJob(FullRunJob.scala:245) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob$.processInBeamCompiler(FullRunJob.scala:140) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob$.main(FullRunJob.scala:56) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob.main(FullRunJob.scala) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2403) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2352) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2351) ~[spark-core_2.12-3.2.0.jar:3.2.0] at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) ~[scala-library-2.12.15.jar:?] at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) ~[scala-library-2.12.15.jar:?] at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) ~[scala-library-2.12.15.jar:?] at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2351) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1109) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1109) ~[spark-core_2.12-3.2.0.jar:3.2.0] at scala.Option.foreach(Option.scala:407) ~[scala-library-2.12.15.jar:?] at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1109) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2591) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2533) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2522) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:898) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2214) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2235) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2254) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2279) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$foreach$1(RDD.scala:1012) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.withScope(RDD.scala:414) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.foreach(RDD.scala:1010) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.api.java.JavaRDDLike.foreach(JavaRDDLike.scala:352) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.api.java.JavaRDDLike.foreach$(JavaRDDLike.scala:351) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.api.java.AbstractJavaRDDLike.foreach(JavaRDDLike.scala:45) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.beam.runners.spark.translation.BoundedDataset.action(BoundedDataset.java:127) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.EvaluationContext.computeOutputs(EvaluationContext.java:228) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkRunner.lambda$run$1(SparkRunner.java:234) ~[beam-runners-spark-3-2.36.0.jar:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?] at java.lang.Thread.run(Thread.java:829) ~[?:?] [2022-11-14T11:06:48,929] [INFO ] o.a.s.SparkContext - Invoking stop() from shutdown hook [2022-11-14T11:06:48,964] [INFO ] o.a.s.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped! [2022-11-14T11:06:48,986] [INFO ] o.t.s.c.c.ContainerManager - Closed container processing-beam [2022-11-14T11:06:48,988] [INFO ] o.a.s.s.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped! [2022-11-14T11:06:49,011] [INFO ] o.t.s.c.c.ContainerManager - Closed container localio [2022-11-14T11:06:49,011] [INFO ] o.t.s.c.r.m.ComponentManager - Released the contextual ComponentManager instance (classloader=jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27, jvm=19385@793948bf7ece) [2022-11-14T11:06:49,015] [INFO ] o.a.s.SparkContext - Successfully stopped SparkContext [2022-11-14T11:06:49,015] [INFO ] o.a.s.u.ShutdownHookManager - Shutdown hook called [2022-11-14T11:06:49,016] [INFO ] o.a.s.u.ShutdownHookManager - Deleting directory /tmp/spark-24e589e7-7efd-4778-87d9-20351eba7c56