Description
Create a simple flow JDBCInput -> HDFSoutput with FlowRunner and run the JOB.
If there is no data in the table it result to an exception :
hdfs://tldns01/user/hellweek/datastreams_backend/output/filtercolumnEF89sfedsddffssdfdzff3dfsfddgffdd1dd/_temporary/0 does not exist.",
"at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:36)",
"at org.apache.beam.sdk.io.Write$Bound$2$auxiliary$H3oVD3p8.invokeProcessElement(Unknown Source)",
"at org.apache.beam.sdk.transforms.DoFnAdapters$SimpleDoFnAdapter.processElement(DoFnAdapters.java:237)",
"at org.apache.beam.runners.spark.translation.SparkProcessContext$ProcCtxtIterator.invokeProcessElement(SparkProcessContext.java:372)",
"at org.apache.beam.runners.spark.translation.SparkProcessContext$ProcCtxtIterator.computeNext(SparkProcessContext.java:335)",
"at com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)",
"at com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)",
"at scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:42)",
"at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:284)",
"at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171)",
"at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78)",
"at org.apache.spark.rdd.RDD.iterator(RDD.scala:268)",
"at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)",
"at org.apache.spark.scheduler.Task.run(Task.scala:89)",
"at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)",
"at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)",
"at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)",
"at java.lang.Thread.run(Thread.java:745)",
"Caused by: java.io.FileNotFoundException: File hdfs://tldns01/user/hellweek/datastreams_backend/output/filtercolumnEF89sfedsddffssdfdzff3dfsfddgffdd1dd/_temporary/0 does not exist.",
"at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:795)",
"at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106)",
"at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853)",
"at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849)",
"at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)",
"at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860)",
"at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)",
"at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1557)",
"at org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.getAllCommittedTaskPaths(FileOutputCommitter.java:291)",
"at org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.commitJobInternal(FileOutputCommitter.java:361)",
"at org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.commitJob(FileOutputCommitter.java:334)",
"at org.apache.beam.sdk.io.hdfs.ConfigurableHDFSFileSink$HDFSWriteOperation.finalize(ConfigurableHDFSFileSink.java:171)",
"at org.apache.beam.sdk.io.Write$Bound$2.processElement(Write.java:417)",
Add Comment Collapse