[2022-11-14T11:08:35,866] [INFO ] o.t.d.s.FullRunJob$ - [ option: runtimeFlow runtimeFlow [ARG] :: Base64 encoded runtimeflow ] [2022-11-14T11:08:35,871] [INFO ] o.t.d.s.FullRunJob$ - [ option: pluginIds pluginIds [ARG] :: Ignored ] [2022-11-14T11:08:35,871] [INFO ] o.t.d.s.FullRunJob$ - [ option: metricsSinkUrl metricsSinkUrl [ARG] :: Metrics sink URL ] [2022-11-14T11:08:35,872] [INFO ] o.t.d.s.FullRunJob$ - [ option: metricsRefreshInterval metricsRefreshInterval [ARG] :: Metrics refresh interval ] [2022-11-14T11:08:35,872] [INFO ] o.t.d.s.FullRunJob$ - Reading RuntimeFlow from runtimeFlow argument [2022-11-14T11:08:35,983] [INFO ] v.JsonUtils$ - Getting Vault key name from the received payload [2022-11-14T11:08:36,036] [INFO ] v.VaultProxy - Client doesn't provide any 'sensitiveFields', it will apply default regex mechanism to decrypt. [2022-11-14T11:08:36,040] [INFO ] v.VaultProxy - The payload doesn't contain any value to be encrypted [2022-11-14T11:08:36,739] [INFO ] o.t.d.s.u.MetricsHelper$ - Using Metrics parameters from Livy environment variables [2022-11-14T11:08:36,812] [INFO ] o.t.d.s.u.ReplicateContentHelper$ - Disabling Beam splittable DoFns for compatibility reasons [2022-11-14T11:08:36,822] [INFO ] o.t.d.s.FullRunJob$ - Done parsing and validating the RuntimeFlow, starting the job... [2022-11-14T11:08:36,996] [INFO ] o.t.s.c.c.ContainerManager - No TALEND-INF/plugins.properties found, will use file resolution [2022-11-14T11:08:37,034] [INFO ] o.t.s.c.r.m.ComponentManager - ComponentManager version: 1.51.0-SNAPSHOT [2022-11-14T11:08:37,039] [INFO ] o.t.s.c.r.m.ComponentManager - Creating the contextual ComponentManager instance (classloader=jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27, jvm=19762@793948bf7ece) [2022-11-14T11:08:37,040] [INFO ] o.t.s.c.r.m.ComponentManager - Components: [] [2022-11-14T11:08:37,041] [INFO ] o.t.s.c.r.m.ComponentManager - Created the contextual ComponentManager instance (classloader=jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27, jvm=19762@793948bf7ece) [2022-11-14T11:08:37,045] [INFO ] o.t.d.t.ConnectorsRepoContainerFinder - Initializing from /opt/talend/connectors [2022-11-14T11:08:37,055] [INFO ] o.t.s.c.c.ContainerManager - Creating module /opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar (from /opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar, location=/opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar) [2022-11-14T11:08:37,361] [INFO ] o.t.s.c.r.m.ComponentManager - Added @Service class org.talend.components.processing.ProcessingService for container-id=processing-beam [2022-11-14T11:08:37,413] [INFO ] o.t.s.c.r.m.ComponentManager - Added @Service class org.talend.components.processing.join.JoinService for container-id=processing-beam [2022-11-14T11:08:37,453] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.aggregate.Aggregate for container-id=processing-beam [2022-11-14T11:08:37,462] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.applyschema.ApplySchema for container-id=processing-beam [2022-11-14T11:08:37,467] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.fieldremover.FieldRemover for container-id=processing-beam [2022-11-14T11:08:37,501] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.fieldselector.FieldSelector for container-id=processing-beam [2022-11-14T11:08:37,526] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.filter.Filter for container-id=processing-beam [2022-11-14T11:08:37,530] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.filter.SemanticFilter for container-id=processing-beam [2022-11-14T11:08:37,534] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.filter.TechnicalFilter for container-id=processing-beam [2022-11-14T11:08:37,539] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.flatten.Flatten for container-id=processing-beam [2022-11-14T11:08:37,552] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.join.JoinJobBuilder for container-id=processing-beam [2022-11-14T11:08:37,556] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.limit.Limit for container-id=processing-beam [2022-11-14T11:08:37,558] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.limit.Sample for container-id=processing-beam [2022-11-14T11:08:37,563] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.python.Python for container-id=processing-beam [2022-11-14T11:08:37,567] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.replicate.Replicate for container-id=processing-beam [2022-11-14T11:08:37,578] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.sampling.Sampling for container-id=processing-beam [2022-11-14T11:08:37,582] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.typeconverter.TypeConverter for container-id=processing-beam [2022-11-14T11:08:37,589] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.processing.window.Window for container-id=processing-beam [2022-11-14T11:08:37,610] [INFO ] o.t.c.e.r.i.RegisteredComponentExtension - RegisteredComponentExtension extension will load CustomComponentExtension... [2022-11-14T11:08:37,612] [INFO ] o.t.s.c.c.ContainerManager - Created container processing-beam [2022-11-14T11:08:37,612] [INFO ] o.t.s.c.r.m.ComponentManager - Adding plugin: /opt/talend/connectors/org/talend/components/processing-beam/1.40.0-SNAPSHOT/processing-beam-1.40.0-SNAPSHOT.jar, as processing-beam [2022-11-14T11:08:37,613] [INFO ] o.t.s.c.c.ContainerManager - Creating module /opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar (from /opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar, location=/opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar) [2022-11-14T11:08:37,728] [INFO ] o.t.s.c.r.m.ComponentManager - Added @Service class org.talend.components.localio.fixed.FixedFlowInputService for container-id=localio [2022-11-14T11:08:37,741] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.localio.fixed.FixedFlowInputRuntime for container-id=localio [2022-11-14T11:08:37,745] [INFO ] o.t.s.c.r.m.ComponentManager - Parsed component class org.talend.components.localio.devnull.DevNullOutputRuntime for container-id=localio [2022-11-14T11:08:37,746] [INFO ] o.t.c.e.r.i.RegisteredComponentExtension - RegisteredComponentExtension extension will load CustomComponentExtension... [2022-11-14T11:08:37,747] [INFO ] o.t.s.c.c.ContainerManager - Created container localio [2022-11-14T11:08:37,747] [INFO ] o.t.s.c.r.m.ComponentManager - Adding plugin: /opt/talend/connectors/org/talend/components/localio/1.40.0-SNAPSHOT/localio-1.40.0-SNAPSHOT.jar, as localio [2022-11-14T11:08:39,167] [INFO ] o.t.d.b.c.r.RuntimeFlowBeamCompilerContext - Creating Beam Pipeline by: class com.sun.proxy.$Proxy104 [2022-11-14T11:08:39,314] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:39,579] [INFO ] o.t.d.b.c.BeamCompiler - Start to resolve runtime for component bG9jYWxpbyNMb2NhbElPI0ZpeGVkRmxvd0lucHV0UnVudGltZQ (id=urk2ljcsypeesuqxkdavzqpt) [2022-11-14T11:08:39,587] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - ====== Get runtime of a component ====== [2022-11-14T11:08:39,611] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - localio#LocalIO#FixedFlowInputRuntime [2022-11-14T11:08:39,614] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Selected a component of type: MAPPER [2022-11-14T11:08:39,619] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Component version is: 1 [2022-11-14T11:08:39,854] [INFO ] o.t.d.b.c.BeamCompiler - Tacokit component runtime: MapperTransform [2022-11-14T11:08:39,921] [WARN ] o.t.s.c.r.s.EnhancedObjectInputStream$Defaults - talend.component.runtime.serialization.java.inputstream.whitelist system property not set, will use default blacklist but this is not considered as a secure setup. Blacklisted packages: [com.sun.rowset., org.apache.ibatis., org.apache.xalan, com.sun.org.apache.xalan.internal.xsltc.trax., java.rmi.server., oracle.jdbc., java.lang.Process, org.springframework.beans.factory.config., org.slf4j.ext., java.util.logging., com.sun.org.apache.bcel.internal.util., com.sun.deploy.security.ruleset., org.apache.axis2.jaxws.spi.handler., org.apache.commons.collections4.functors., org.apache.commons.collections.functors., org.hibernate.jmx., flex.messaging.util.concurrent., org.apache.openjpa.ee., org.codehaus.groovy.runtime., org.apache.axis2.transport.jms., org.apache.tomcat.dbcp.dbcp2., jodd.db.connection., org.jboss.util.propertyeditor.] [2022-11-14T11:08:40,317] [INFO ] o.t.d.b.c.BeamCompiler - Start to resolve runtime for component cHJvY2Vzc2luZy1iZWFtI1Byb2Nlc3NpbmcjUHl0aG9u (id=t0gf8f3bo6eqvujwx8b9irts) [2022-11-14T11:08:40,321] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - ====== Get runtime of a component ====== [2022-11-14T11:08:40,321] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - processing-beam#Processing#Python [2022-11-14T11:08:40,322] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Selected a component of type: PROCESSOR [2022-11-14T11:08:40,325] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Component version is: 1 [2022-11-14T11:08:40,337] [WARN ] o.t.s.c.r.b.s.BeamComponentExtension - Creating a 'interface org.talend.sdk.component.runtime.output.Processor' instance for 'Processing#Python', this must be unwrapped before being used [2022-11-14T11:08:40,343] [WARN ] o.t.s.c.r.b.s.BeamComponentExtension - Creating a 'interface org.talend.sdk.component.runtime.output.Processor' instance for 'Processing#Python', this must be unwrapped before being used [2022-11-14T11:08:40,345] [INFO ] o.t.d.b.c.BeamCompiler - Tacokit component runtime: Python [2022-11-14T11:08:40,399] [INFO ] o.t.d.b.c.BeamCompiler - Start to resolve runtime for component bG9jYWxpbyNMb2NhbElPI0Rldk51bGxPdXRwdXRSdW50aW1l (id=buezmtcmhxk3jmifu1zptbro) [2022-11-14T11:08:40,403] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - ====== Get runtime of a component ====== [2022-11-14T11:08:40,404] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - localio#LocalIO#DevNullOutputRuntime [2022-11-14T11:08:40,404] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Selected a component of type: PROCESSOR [2022-11-14T11:08:40,405] [INFO ] o.t.d.b.c.r.GenericRuntimeBuilder - Component version is: 1 [2022-11-14T11:08:40,439] [INFO ] o.t.d.b.c.BeamCompiler - Tacokit component runtime: org.talend.sdk.component.runtime.output.ProcessorImpl@bcd4fffe [2022-11-14T11:08:40,470] [WARN ] o.t.d.b.c.r.RuntimeFlowBeamCompilerContext - Not able to parse 'metrics.socketTimeoutMs', default is 15000 [2022-11-14T11:08:40,471] [WARN ] o.t.d.b.c.r.RuntimeFlowBeamCompilerContext - Not able to parse 'metrics.connectionTimeoutMs', default is 15000 [2022-11-14T11:08:40,714] [INFO ] o.a.s.SparkContext - Running Spark version 3.2.0 [2022-11-14T11:08:40,910] [WARN ] o.a.h.u.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable [2022-11-14T11:08:41,023] [INFO ] o.a.s.r.ResourceUtils - ============================================================== [2022-11-14T11:08:41,024] [INFO ] o.a.s.r.ResourceUtils - No custom resources configured for spark.driver. [2022-11-14T11:08:41,024] [INFO ] o.a.s.r.ResourceUtils - ============================================================== [2022-11-14T11:08:41,025] [INFO ] o.a.s.SparkContext - Submitted application: pipeline for proc-python2-subprocess processor [2022-11-14T11:08:41,045] [INFO ] o.a.s.r.ResourceProfile - Default ResourceProfile created, executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , memory -> name: memory, amount: 1024, script: , vendor: , offHeap -> name: offHeap, amount: 0, script: , vendor: ), task resources: Map(cpus -> name: cpus, amount: 1.0) [2022-11-14T11:08:41,051] [INFO ] o.a.s.r.ResourceProfile - Limiting resource is cpu [2022-11-14T11:08:41,052] [INFO ] o.a.s.r.ResourceProfileManager - Added ResourceProfile id: 0 [2022-11-14T11:08:41,108] [INFO ] o.a.s.SecurityManager - Changing view acls to: root [2022-11-14T11:08:41,108] [INFO ] o.a.s.SecurityManager - Changing modify acls to: root [2022-11-14T11:08:41,109] [INFO ] o.a.s.SecurityManager - Changing view acls groups to: [2022-11-14T11:08:41,109] [INFO ] o.a.s.SecurityManager - Changing modify acls groups to: [2022-11-14T11:08:41,110] [INFO ] o.a.s.SecurityManager - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); groups with view permissions: Set(); users with modify permissions: Set(root); groups with modify permissions: Set() [2022-11-14T11:08:41,396] [INFO ] o.a.s.SparkEnv - Registering MapOutputTracker [2022-11-14T11:08:41,426] [INFO ] o.a.s.SparkEnv - Registering BlockManagerMaster [2022-11-14T11:08:41,448] [INFO ] o.a.s.SparkEnv - Registering BlockManagerMasterHeartbeat [2022-11-14T11:08:41,518] [INFO ] o.a.s.SparkEnv - Registering OutputCommitCoordinator [2022-11-14T11:08:41,627] [INFO ] o.a.s.e.Executor - Starting executor ID driver on host 793948bf7ece [2022-11-14T11:08:41,715] [INFO ] o.s.j.u.log - Logging initialized @7283ms to org.sparkproject.jetty.util.log.Slf4jLog [2022-11-14T11:08:41,786] [INFO ] o.t.d.b.c.BeamJob - Context variables initialization for this execution: [2022-11-14T11:08:41,806] [INFO ] o.t.c.c.d.p.DictionaryProvider - Loading registry from /tmp/dict/12e4d32a-69ea-4956-b382-dbb5704da353/indexFolder [2022-11-14T11:08:41,809] [INFO ] o.t.c.c.d.p.DictionaryProvider - The dictionary '0' can't be found in the local registry - creating it [2022-11-14T11:08:41,844] [INFO ] o.t.d.s.a.CategoryRegistryManager - Loading categories from local registry. [2022-11-14T11:08:42,018] [INFO ] o.t.d.s.i.ClassPathDirectory - Opening 'file:///tmp/dict/indexFolder/shared/prod/dictionary/' ... [2022-11-14T11:08:42,019] [INFO ] o.t.d.s.i.ClassPathDirectory - Opening 'file:///tmp/dict/indexFolder/shared/prod/keyword/' ... [2022-11-14T11:08:42,318] [INFO ] o.t.d.s.FullRunJob$ - Got dictionary 0 from MavenDictionaryProvider. [2022-11-14T11:08:42,318] [WARN ] o.t.d.s.FullRunJob$ - DQ rules is disabled for now, pending TPRUN-4490 [2022-11-14T11:08:42,322] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:42,327] [WARN ] o.a.b.s.Pipeline - The following transforms do not have stable unique names: , ParDo(VolumeOnLink), ParDo(NumberRecordsOnLink) [2022-11-14T11:08:42,327] [INFO ] o.a.b.r.s.SparkRunner - Executing pipeline using the SparkRunner. [2022-11-14T11:08:42,613] [INFO ] o.a.b.r.s.t.SparkContextFactory - Using a provided Spark Context [2022-11-14T11:08:42,810] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:42,843] [INFO ] o.a.b.r.s.m.MetricsAccumulator - Instantiated metrics accumulator: MetricQueryResults() [2022-11-14T11:08:42,845] [INFO ] o.a.b.r.s.a.AggregatorsAccumulator - Instantiated aggregators accumulator: [2022-11-14T11:08:42,846] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating Read(BoundedSourceImpl) [2022-11-14T11:08:42,907] [INFO ] o.t.d.b.c.BeamJob - Metrics are enabled [2022-11-14T11:08:42,910] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.util.builder.tacokit.transform.RecordToIndexedRecord@31773d5b [2022-11-14T11:08:43,105] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$NumberRecordsOnLink@71dac63f [2022-11-14T11:08:43,123] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$VolumeOnLink@5db6083b [2022-11-14T11:08:43,136] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.components.processing.python.PythonDoFn@c11ec9b [2022-11-14T11:08:43,153] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$NumberRecordsOnLink@118cbb26 [2022-11-14T11:08:43,165] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.metrics.CountersOnLinks$VolumeOnLink@36a59d8 [2022-11-14T11:08:43,177] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.datastreams.beam.compiler.util.builder.tacokit.transform.IndexedRecordToRecord@3299e315 [2022-11-14T11:08:43,190] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.sdk.component.runtime.beam.transform.RecordNormalizer@10007fdc [2022-11-14T11:08:43,203] [INFO ] o.a.b.r.s.SparkRunner$Evaluator - Evaluating org.talend.sdk.component.runtime.beam.TalendIO$WriteFn@699fe19f [2022-11-14T11:08:43,265] [INFO ] o.a.s.SparkContext - Starting job: foreach at BoundedDataset.java:127 [2022-11-14T11:08:43,288] [INFO ] o.a.s.s.DAGScheduler - Got job 0 (foreach at BoundedDataset.java:127) with 5 output partitions [2022-11-14T11:08:43,288] [INFO ] o.a.s.s.DAGScheduler - Final stage: ResultStage 0 (foreach at BoundedDataset.java:127) [2022-11-14T11:08:43,289] [INFO ] o.a.s.s.DAGScheduler - Parents of final stage: List() [2022-11-14T11:08:43,292] [INFO ] o.a.s.s.DAGScheduler - Missing parents: List() [2022-11-14T11:08:43,298] [INFO ] o.a.s.s.DAGScheduler - Submitting ResultStage 0 (3/Processor_LocalIO#DevNullOutputRuntime/ParDo(Write)/ParMultiDo(Write).output MapPartitionsRDD[27] at values at TransformTranslator.java:442), which has no missing parents [2022-11-14T11:08:43,480] [INFO ] o.a.s.SparkContext - Created broadcast 0 from broadcast at DAGScheduler.scala:1427 [2022-11-14T11:08:43,491] [INFO ] o.a.s.s.DAGScheduler - Submitting 5 missing tasks from ResultStage 0 (3/Processor_LocalIO#DevNullOutputRuntime/ParDo(Write)/ParMultiDo(Write).output MapPartitionsRDD[27] at values at TransformTranslator.java:442) (first 15 tasks are for partitions Vector(0, 1, 2, 3, 4)) [2022-11-14T11:08:43,493] [INFO ] o.a.s.s.TaskSchedulerImpl - Adding task set 0.0 with 5 tasks resource profile 0 [2022-11-14T11:08:43,561] [INFO ] o.a.s.s.TaskSetManager - Starting task 0.0 in stage 0.0 (TID 0) (793948bf7ece, executor driver, partition 0, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:08:43,572] [INFO ] o.a.s.s.TaskSetManager - Starting task 1.0 in stage 0.0 (TID 1) (793948bf7ece, executor driver, partition 1, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:08:43,574] [INFO ] o.a.s.s.TaskSetManager - Starting task 2.0 in stage 0.0 (TID 2) (793948bf7ece, executor driver, partition 2, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:08:43,575] [INFO ] o.a.s.s.TaskSetManager - Starting task 3.0 in stage 0.0 (TID 3) (793948bf7ece, executor driver, partition 3, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:08:43,592] [INFO ] o.a.s.e.Executor - Running task 0.0 in stage 0.0 (TID 0) [2022-11-14T11:08:43,595] [INFO ] o.a.s.e.Executor - Running task 2.0 in stage 0.0 (TID 2) [2022-11-14T11:08:43,595] [INFO ] o.a.s.e.Executor - Running task 1.0 in stage 0.0 (TID 1) [2022-11-14T11:08:43,597] [INFO ] o.a.s.e.Executor - Running task 3.0 in stage 0.0 (TID 3) [2022-11-14T11:08:43,711] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:43,718] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:43,720] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:43,725] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:44,641] [INFO ] o.a.s.e.Executor - Finished task 1.0 in stage 0.0 (TID 1). 7981 bytes result sent to driver [2022-11-14T11:08:44,642] [INFO ] o.a.s.e.Executor - Finished task 2.0 in stage 0.0 (TID 2). 8024 bytes result sent to driver [2022-11-14T11:08:44,651] [INFO ] o.a.s.e.Executor - Finished task 0.0 in stage 0.0 (TID 0). 7981 bytes result sent to driver [2022-11-14T11:08:44,652] [INFO ] o.a.s.e.Executor - Finished task 3.0 in stage 0.0 (TID 3). 7981 bytes result sent to driver [2022-11-14T11:08:44,653] [INFO ] o.a.s.s.TaskSetManager - Starting task 4.0 in stage 0.0 (TID 4) (793948bf7ece, executor driver, partition 4, PROCESS_LOCAL, 6756 bytes) taskResourceAssignments Map() [2022-11-14T11:08:44,654] [INFO ] o.a.s.e.Executor - Running task 4.0 in stage 0.0 (TID 4) [2022-11-14T11:08:44,669] [INFO ] o.a.s.s.TaskSetManager - Finished task 1.0 in stage 0.0 (TID 1) in 1096 ms on 793948bf7ece (executor driver) (1/5) [2022-11-14T11:08:44,673] [INFO ] o.a.s.s.TaskSetManager - Finished task 3.0 in stage 0.0 (TID 3) in 1097 ms on 793948bf7ece (executor driver) (2/5) [2022-11-14T11:08:44,676] [INFO ] o.a.s.s.TaskSetManager - Finished task 2.0 in stage 0.0 (TID 2) in 1103 ms on 793948bf7ece (executor driver) (3/5) [2022-11-14T11:08:44,676] [INFO ] o.a.s.s.TaskSetManager - Finished task 0.0 in stage 0.0 (TID 0) in 1127 ms on 793948bf7ece (executor driver) (4/5) [2022-11-14T11:08:44,677] [INFO ] o.a.b.s.i.a.s.DefaultS3ClientBuilderFactory - The AWS S3 Beam extension was included in this build, but the awsRegion flag was not specified. If you don't plan to use S3, then ignore this message. [2022-11-14T11:08:49,288] [ERROR] o.a.s.e.Executor - Exception in task 4.0 in stage 0.0 (TID 4) org.apache.beam.sdk.util.UserCodeException: org.talend.components.processing.ProcessingErrorCode$TalendMsgRuntimeException: The import 'subprocess' cannot be used. at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.talend.components.processing.python.PythonDoFn$DoFnInvoker.invokeSetup(Unknown Source) ~[?:?] at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:53) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:129) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:63) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsToPair$1(JavaRDDLike.scala:186) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$mapPartitions$2(RDD.scala:863) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$mapPartitions$2$adapted(RDD.scala:863) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.Task.run(Task.scala:131) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) [spark-core_2.12-3.2.0.jar:3.2.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] Suppressed: org.apache.beam.sdk.util.UserCodeException: java.lang.NullPointerException at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.talend.components.processing.python.PythonDoFn$DoFnInvoker.invokeTeardown(Unknown Source) ~[?:?] at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:56) ~[beam-sdks-java-harness-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:129) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:63) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsToPair$1(JavaRDDLike.scala:186) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$mapPartitions$2(RDD.scala:863) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$mapPartitions$2$adapted(RDD.scala:863) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.iterator(RDD.scala:337) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.Task.run(Task.scala:131) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509) [spark-core_2.12-3.2.0.jar:3.2.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] Caused by: java.lang.NullPointerException at org.talend.components.processing.python.PythonDoFn.tearDown(PythonDoFn.java:240) ~[?:?] Caused by: org.talend.components.processing.ProcessingErrorCode$TalendMsgRuntimeException: The import 'subprocess' cannot be used. at org.talend.components.processing.ProcessingErrorCode.createInvalidPythonImportErrorException(ProcessingErrorCode.java:137) ~[?:?] at org.talend.components.processing.python.PythonDoFn$1.visitImport(PythonDoFn.java:96) ~[?:?] at org.python.antlr.ast.Import.accept(Import.java:133) ~[?:?] at org.python.antlr.ast.TryExcept.traverse(TryExcept.java:221) ~[?:?] at org.python.antlr.Visitor.traverse(Visitor.java:12) ~[?:?] at org.python.antlr.ast.VisitorBase.visitTryExcept(VisitorBase.java:126) ~[?:?] at org.python.antlr.ast.TryExcept.accept(TryExcept.java:214) ~[?:?] at org.python.antlr.ast.FunctionDef.traverse(FunctionDef.java:228) ~[?:?] at org.python.antlr.Visitor.traverse(Visitor.java:12) ~[?:?] at org.python.antlr.ast.VisitorBase.visitFunctionDef(VisitorBase.java:54) ~[?:?] at org.python.antlr.ast.FunctionDef.accept(FunctionDef.java:219) ~[?:?] at org.python.antlr.Visitor.visit(Visitor.java:26) ~[?:?] at org.talend.components.processing.python.PythonDoFn.checkImportBlacklist(PythonDoFn.java:126) ~[?:?] at org.talend.components.processing.python.PythonDoFn.setup(PythonDoFn.java:64) ~[?:?] [2022-11-14T11:08:49,347] [WARN ] o.a.s.ThrowableSerializationWrapper - Task exception could not be deserialized java.lang.ClassNotFoundException: org.talend.components.processing.ProcessingErrorCode$TalendMsgRuntimeException at jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581) ~[?:?] at jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) ~[?:?] at java.lang.ClassLoader.loadClass(ClassLoader.java:522) ~[?:?] at java.lang.Class.forName0(Native Method) ~[?:?] at java.lang.Class.forName(Class.java:398) ~[?:?] at org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:68) ~[spark-core_2.12-3.2.0.jar:3.2.0] at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1986) ~[?:?] at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1853) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2184) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2479) ~[?:?] at java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:624) ~[?:?] at java.lang.Throwable.readObject(Throwable.java:896) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1046) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2340) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:488) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:446) ~[?:?] at org.apache.spark.ThrowableSerializationWrapper.readObject(TaskEndReason.scala:202) ~[spark-core_2.12-3.2.0.jar:3.2.0] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1046) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2340) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2479) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2373) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2479) ~[?:?] at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2373) ~[?:?] at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2211) ~[?:?] at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1670) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:488) ~[?:?] at java.io.ObjectInputStream.readObject(ObjectInputStream.java:446) ~[?:?] at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:76) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:115) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.TaskResultGetter.$anonfun$enqueueFailedTask$2(TaskResultGetter.scala:141) ~[spark-core_2.12-3.2.0.jar:3.2.0] at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) ~[scala-library-2.12.15.jar:?] at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.TaskResultGetter.$anonfun$enqueueFailedTask$1(TaskResultGetter.scala:137) ~[spark-core_2.12-3.2.0.jar:3.2.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] [2022-11-14T11:08:49,349] [ERROR] o.a.s.s.TaskResultGetter - Could not deserialize TaskEndReason: ClassNotFound with classloader jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27 [2022-11-14T11:08:49,350] [WARN ] o.a.s.s.TaskSetManager - Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason [2022-11-14T11:08:49,352] [ERROR] o.a.s.s.TaskSetManager - Task 4 in stage 0.0 failed 1 times; aborting job [2022-11-14T11:08:49,353] [INFO ] o.a.s.s.TaskSchedulerImpl - Removed TaskSet 0.0, whose tasks have all completed, from pool [2022-11-14T11:08:49,355] [INFO ] o.a.s.s.TaskSchedulerImpl - Cancelling stage 0 [2022-11-14T11:08:49,355] [INFO ] o.a.s.s.TaskSchedulerImpl - Killing all running tasks in stage 0: Stage cancelled [2022-11-14T11:08:49,357] [INFO ] o.a.s.s.DAGScheduler - ResultStage 0 (foreach at BoundedDataset.java:127) failed in 6.020 s due to Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: [2022-11-14T11:08:49,362] [INFO ] o.a.s.s.DAGScheduler - Job 0 failed: foreach at BoundedDataset.java:127, took 6.096384 s [2022-11-14T11:08:49,364] [INFO ] o.t.d.b.c.r.ServerSocketAvroSchemaRegistry - Shutdown registry containing {} [2022-11-14T11:08:49,365] [ERROR] o.t.d.s.FullRunJob$ - DATA_STREAMS_ERROR_MSG:Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: [2022-11-14T11:08:49,365] [ERROR] o.t.d.s.FullRunJob$ - Original Throwable: java.lang.RuntimeException: org.apache.spark.SparkException: Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: at org.apache.beam.runners.spark.SparkPipelineResult.runtimeExceptionFrom(SparkPipelineResult.java:60) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkPipelineResult.beamExceptionFrom(SparkPipelineResult.java:77) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:104) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:92) ~[beam-runners-spark-3-2.36.0.jar:?] at org.talend.datastreams.beam.compiler.BeamJob.waitUntilFinish(BeamJob.java:211) ~[data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.beam.compiler.BeamJob.fullRun(BeamJob.java:180) ~[data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob$.runJob(FullRunJob.scala:245) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob$.processInBeamCompiler(FullRunJob.scala:140) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob$.main(FullRunJob.scala:56) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] at org.talend.datastreams.streamsjob.FullRunJob.main(FullRunJob.scala) [data-processing-runtime-streamsjob-2.21.0-PR-924-SNAPSHOT.jar:2.21.0-PR-924-SNAPSHOT] Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 4 in stage 0.0 failed 1 times, most recent failure: Lost task 4.0 in stage 0.0 (TID 4) (793948bf7ece executor driver): UnknownReason Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2403) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2352) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2351) ~[spark-core_2.12-3.2.0.jar:3.2.0] at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) ~[scala-library-2.12.15.jar:?] at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) ~[scala-library-2.12.15.jar:?] at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) ~[scala-library-2.12.15.jar:?] at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2351) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1109) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1109) ~[spark-core_2.12-3.2.0.jar:3.2.0] at scala.Option.foreach(Option.scala:407) ~[scala-library-2.12.15.jar:?] at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1109) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2591) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2533) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2522) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:898) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2214) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2235) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2254) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.SparkContext.runJob(SparkContext.scala:2279) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.$anonfun$foreach$1(RDD.scala:1012) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.withScope(RDD.scala:414) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.rdd.RDD.foreach(RDD.scala:1010) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.api.java.JavaRDDLike.foreach(JavaRDDLike.scala:352) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.api.java.JavaRDDLike.foreach$(JavaRDDLike.scala:351) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.spark.api.java.AbstractJavaRDDLike.foreach(JavaRDDLike.scala:45) ~[spark-core_2.12-3.2.0.jar:3.2.0] at org.apache.beam.runners.spark.translation.BoundedDataset.action(BoundedDataset.java:127) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.translation.EvaluationContext.computeOutputs(EvaluationContext.java:228) ~[beam-runners-spark-3-2.36.0.jar:?] at org.apache.beam.runners.spark.SparkRunner.lambda$run$1(SparkRunner.java:234) ~[beam-runners-spark-3-2.36.0.jar:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?] at java.lang.Thread.run(Thread.java:829) ~[?:?] [2022-11-14T11:08:49,375] [INFO ] o.a.s.SparkContext - Invoking stop() from shutdown hook [2022-11-14T11:08:49,398] [INFO ] o.a.s.MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped! [2022-11-14T11:08:49,433] [INFO ] o.a.s.s.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped! [2022-11-14T11:08:49,438] [INFO ] o.t.s.c.c.ContainerManager - Closed container processing-beam [2022-11-14T11:08:49,442] [INFO ] o.a.s.SparkContext - Successfully stopped SparkContext [2022-11-14T11:08:49,444] [INFO ] o.a.s.u.ShutdownHookManager - Shutdown hook called [2022-11-14T11:08:49,447] [INFO ] o.a.s.u.ShutdownHookManager - Deleting directory /tmp/spark-5c45b265-178d-4be1-bd28-e367e98935ff [2022-11-14T11:08:49,448] [INFO ] o.t.s.c.c.ContainerManager - Closed container localio [2022-11-14T11:08:49,448] [INFO ] o.t.s.c.r.m.ComponentManager - Released the contextual ComponentManager instance (classloader=jdk.internal.loader.ClassLoaders$AppClassLoader@5ffd2b27, jvm=19762@793948bf7ece)