I am trying to read sas files with tihs package . With smaller file size it is working fine , But filesize greater then 512 MB it is giving me below error
ERROR InsertIntoHadoopFsRelationCommand: Aborting job.
org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 8, 10.0.0.5): org.apache.spark.SparkException: Task failed while writing rows
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:261)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.github.saurfang.sas.util.PrivateMethodCaller.apply(PrivateMethodExposer.scala:11)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$lzycompute$1(SasRecordReader.scala:119)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$1(SasRecordReader.scala:118)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:131)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:19)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:254)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply$mcV$sp(WriterContainer.scala:253)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1348)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:258)
... 8 more
Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
at java.util.ArrayList.rangeCheck(ArrayList.java:653)
at java.util.ArrayList.get(ArrayList.java:429)
at com.ggasoftware.parso.SasFileParser.readNext(SasFileParser.java:876)
... 32 more
17/04/04 10:47:32 INFO TaskSetManager: Starting task 3.1 in stage 0.0 (TID 4, 10.0.0.5, partition 3, PROCESS_LOCAL, 5573 bytes)
17/04/04 10:47:32 INFO YarnSchedulerBackend$YarnDriverEndpoint: Launching task 4 on executor id: 1 hostname: 10.0.0.5.
17/04/04 10:47:32 WARN TaskSetManager: Lost task 2.0 in stage 0.0 (TID 2, 10.0.0.5): java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.github.saurfang.sas.util.PrivateMethodCaller.apply(PrivateMethodExposer.scala:11)
at com.github.saurfang.sas.mapred.SasRecordReader.(SasRecordReader.scala:111)
at com.github.saurfang.sas.mapred.SasInputFormat.getRecordReader(SasInputFormat.scala:15)
at org.apache.spark.rdd.HadoopRDD$$anon$1.(HadoopRDD.scala:245)
at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.StackOverflowError
at com.ggasoftware.parso.SasFileParser.readPageHeader(SasFileParser.java:949)
at com.ggasoftware.parso.SasFileParser.readNextPage(SasFileParser.java:928)
.java:935)
at com.ggasoftware.parso.SasFileParser.readNextPage(SasFileParser.java:935)
17/04/04 10:47:32 INFO TaskSetManager: Starting task 2.1 in stage 0.0 (TID 5, 10.0.0.5, partition 2, PROCESS_LOCAL, 5573 bytes)
17/04/04 10:47:32 INFO YarnSchedulerBackend$YarnDriverEndpoint: Launching task 5 on executor id: 2 hostname: 10.0.0.5.
17/04/04 10:47:33 INFO TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1) on executor 10.0.0.5: java.lang.reflect.InvocationTargetException (null) [duplicate 1]
17/04/04 10:47:33 INFO TaskSetManager: Starting task 1.1 in stage 0.0 (TID 6, 10.0.0.5, partition 1, PROCESS_LOCAL, 5573 bytes)
17/04/04 10:47:33 INFO YarnSchedulerBackend$YarnDriverEndpoint: Launching task 6 on executor id: 1 hostname: 10.0.0.5.
17/04/04 10:47:33 INFO TaskSetManager: Lost task 3.1 in stage 0.0 (TID 4) on executor 10.0.0.5: org.apache.spark.SparkException (Task failed while writing rows) [duplicate 1]
17/04/04 10:47:33 INFO TaskSetManager: Starting task 3.2 in stage 0.0 (TID 7, 10.0.0.5, partition 3, PROCESS_LOCAL, 5573 bytes)
17/04/04 10:47:33 INFO YarnSchedulerBackend$YarnDriverEndpoint: Launching task 7 on executor id: 1 hostname: 10.0.0.5.
17/04/04 10:47:34 INFO TaskSetManager: Lost task 3.2 in stage 0.0 (TID 7) on executor 10.0.0.5: org.apache.spark.SparkException (Task failed while writing rows) [duplicate 2]
17/04/04 10:47:34 INFO TaskSetManager: Starting task 3.3 in stage 0.0 (TID 8, 10.0.0.5, partition 3, PROCESS_LOCAL, 5573 bytes)
17/04/04 10:47:34 INFO YarnSchedulerBackend$YarnDriverEndpoint: Launching task 8 on executor id: 1 hostname: 10.0.0.5.
17/04/04 10:47:35 INFO TaskSetManager: Lost task 3.3 in stage 0.0 (TID 8) on executor 10.0.0.5: org.apache.spark.SparkException (Task failed while writing rows) [duplicate 3]
17/04/04 10:47:35 ERROR TaskSetManager: Task 3 in stage 0.0 failed 4 times; aborting job
17/04/04 10:47:35 INFO YarnScheduler: Cancelling stage 0
17/04/04 10:47:35 INFO YarnScheduler: Stage 0 was cancelled
17/04/04 10:47:35 INFO DAGScheduler: ResultStage 0 (save at Wordcount.scala:52) failed in 8.345 s
17/04/04 10:47:35 INFO DAGScheduler: Job 0 failed: save at Wordcount.scala:52, took 8.458114 s
17/04/04 10:47:35 ERROR InsertIntoHadoopFsRelationCommand: Aborting job.
org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 8, 10.0.0.5): org.apache.spark.SparkException: Task failed while writing rows
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:261)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.github.saurfang.sas.util.PrivateMethodCaller.apply(PrivateMethodExposer.scala:11)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$lzycompute$1(SasRecordReader.scala:119)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$1(SasRecordReader.scala:118)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:131)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:19)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:254)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply$mcV$sp(WriterContainer.scala:253)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1348)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:258)
... 8 more
Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
at java.util.ArrayList.rangeCheck(ArrayList.java:653)
at java.util.ArrayList.get(ArrayList.java:429)
at com.ggasoftware.parso.SasFileParser.readNext(SasFileParser.java:876)
... 32 more
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1454)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1442)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1441)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1441)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:811)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:811)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:811)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1667)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1622)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1611)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:632)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1873)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1886)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1906)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:57)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:115)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114)
at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:86)
at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:86)
at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:525)
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211)
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:194)
at com.dataflair.spark.Wordcount$.main(Wordcount.scala:52)
at com.dataflair.spark.Wordcount.main(Wordcount.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:736)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:185)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:210)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: org.apache.spark.SparkException: Task failed while writing rows
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:261)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.github.saurfang.sas.util.PrivateMethodCaller.apply(PrivateMethodExposer.scala:11)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$lzycompute$1(SasRecordReader.scala:119)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$1(SasRecordReader.scala:118)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:131)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:19)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:254)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply$mcV$sp(WriterContainer.scala:253)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1348)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:258)
... 8 more
Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
at java.util.ArrayList.rangeCheck(ArrayList.java:653)
at java.util.ArrayList.get(ArrayList.java:429)
at com.ggasoftware.parso.SasFileParser.readNext(SasFileParser.java:876)
... 32 more
17/04/04 10:47:35 WARN AzureFileSystemThreadPoolExecutor: Disabling threads for Delete operation as thread count 0 is <= 1
17/04/04 10:47:35 ERROR AzureNativeFileSystemStore: Encountered Storage Exception for delete on Blob: Exception Details: The specified blob does not exist. Error Code: BlobNotFound
17/04/04 10:47:35 WARN AzureFileSystemThreadPoolExecutor: Failed to Delete file
:rwxrwxrwx]
17/04/04 10:47:35 ERROR NativeAzureFileSystem: Failed to delete files / subfolders in blob
test/sourceoutput/epei.csv/_temporary
17/04/04 10:47:35 ERROR DefaultWriterContainer: Job job_201704041047_0000 aborted.
Exception in thread "main" org.apache.spark.SparkException: Job aborted.
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelationCommand.scala:149)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:57)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:115)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114)
at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:86)
at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:86)
at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:525)
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211)
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:194)
at com.dataflair.spark.Wordcount$.main(Wordcount.scala:52)
at com.dataflair.spark.Wordcount.main(Wordcount.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:736)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:185)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:210)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 8, 10.0.0.5): org.apache.spark.SparkException: Task failed while writing rows
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:261)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.github.saurfang.sas.util.PrivateMethodCaller.apply(PrivateMethodExposer.scala:11)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$lzycompute$1(SasRecordReader.scala:119)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$1(SasRecordReader.scala:118)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:131)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:19)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:254)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply$mcV$sp(WriterContainer.scala:253)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1348)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:258)
... 8 more
Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
at java.util.ArrayList.rangeCheck(ArrayList.java:653)
at java.util.ArrayList.get(ArrayList.java:429)
at com.ggasoftware.parso.SasFileParser.readNext(SasFileParser.java:876)
... 32 more
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1454)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1442)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1441)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1441)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:811)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:811)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:811)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1667)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1622)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1611)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:632)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1873)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1886)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1906)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelationCommand.scala:143)
... 29 more
Caused by: org.apache.spark.SparkException: Task failed while writing rows
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:261)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.github.saurfang.sas.util.PrivateMethodCaller.apply(PrivateMethodExposer.scala:11)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$lzycompute$1(SasRecordReader.scala:119)
at com.github.saurfang.sas.mapred.SasRecordReader.readNext$1(SasRecordReader.scala:118)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:131)
at com.github.saurfang.sas.mapred.SasRecordReader.next(SasRecordReader.scala:19)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:254)
at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply$mcV$sp(WriterContainer.scala:253)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1348)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:258)
... 8 more
Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
at java.util.ArrayList.rangeCheck(ArrayList.java:653)
at java.util.ArrayList.get(ArrayList.java:429)
at com.ggasoftware.parso.SasFileParser.readNext(SasFileParser.java:876)
... 32 more
17/04/04 10:47:35 INFO SparkContext: Invoking stop() from shutdown hook
17/04/04 10:47:35 INFO ServerConnector: Stopped ServerConnector@4108fa66{HTTP/1.1}{0.0.0.0:4040}
17/04/04 10:47:35 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@27e0f2f5{/stages/stage/kill,null,UNAVAILABLE}
17/04/04 10:47:35 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@9cd25ff{/api,null,UNAVAILABLE}
17/04/04 10:47:35 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@69f63d95{/,null,UNAVAILABLE}
17/04/04 10:47:35 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@660e9100{/static,null,UNAVAILABLE}
17/04/04 10:47:35 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@6928f576{/executors/threadDump/json,null,UNAVAILABLE}
17/04/04 10:47:35 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@182f1e9a{/executors/threadDump,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@758f4f03{/executors/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@61edc883{/executors,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@5cc5b667{/environment/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@b5cc23a{/environment,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@5398edd0{/storage/rdd/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@3591009c{/storage/rdd,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@4152d38d{/storage/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@7555b920{/storage,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@4cc547a{/stages/pool/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@7a11c4c7{/stages/pool,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@7e094740{/stages/stage/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@64c2b546{/stages/stage,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@578524c3{/stages/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@551a20d6{/stages,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@5fe8b721{/jobs/job/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@560cbf1a{/jobs/job,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@740abb5{/jobs/json,null,UNAVAILABLE}
17/04/04 10:47:36 INFO ContextHandler: Stopped o.s.j.s.ServletContextHandler@56db847e{/jobs,null,UNAVAILABLE}
17/04/04 10:47:36 INFO SparkUI: Stopped Spark web UI at http://10.0.0.17:4040
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.BlockManager.disk.diskSpaceUsed_MB, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.BlockManager.memory.maxMem_MB, value=8744
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.BlockManager.memory.memUsed_MB, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.BlockManager.memory.remainingMem_MB, value=8744
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.DAGScheduler.job.activeJobs, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.DAGScheduler.job.allJobs, value=1
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.DAGScheduler.stage.failedStages, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.DAGScheduler.stage.runningStages, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.DAGScheduler.stage.waitingStages, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.PS-MarkSweep.count, value=3
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.PS-MarkSweep.time, value=195
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.PS-Scavenge.count, value=6
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.PS-Scavenge.time, value=120
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.heap.committed, value=686292992
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.heap.init, value=461373440
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.heap.max, value=954728448
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.heap.usage, value=0.18328408288929504
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.heap.used, value=175308568
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.non-heap.committed, value=104161280
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.non-heap.init, value=2555904
17/04/04 10:47:36 INFO TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0) on executor 10.0.0.5: org.apache.spark.SparkException (Task failed while writing rows) [duplicate 4]
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.non-heap.max, value=-1
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.non-heap.usage, value=-1.0187864E8
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.non-heap.used, value=101881368
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Code-Cache.committed, value=17170432
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Code-Cache.init, value=2555904
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Code-Cache.max, value=251658240
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Code-Cache.usage, value=0.06387176513671874
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Code-Cache.used, value=16073856
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Compressed-Class-Space.committed, value=10747904
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Compressed-Class-Space.init, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Compressed-Class-Space.max, value=1073741824
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Compressed-Class-Space.usage, value=0.009773895144462585
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Compressed-Class-Space.used, value=10494640
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Metaspace.committed, value=76242944
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Metaspace.init, value=0
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Metaspace.max, value=-1
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Metaspace.usage, value=0.9880171468719781
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.Metaspace.used, value=75329336
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Eden-Space.committed, value=199229440
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Eden-Space.init, value=115867648
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Eden-Space.max, value=259522560
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Eden-Space.usage, value=0.06554482199928978
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Eden-Space.used, value=17010360
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Old-Gen.committed, value=468189184
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Old-Gen.init, value=307757056
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Old-Gen.max, value=716177408
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Old-Gen.usage, value=0.1961920474319123
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Old-Gen.used, value=140508312
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Survivor-Space.committed, value=18874368
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Survivor-Space.init, value=18874368
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Survivor-Space.max, value=18874368
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Survivor-Space.usage, value=0.9998643663194444
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.pools.PS-Survivor-Space.used, value=18871808
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.total.committed, value=790454272
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.total.init, value=463929344
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.total.max, value=954728447
17/04/04 10:47:36 INFO metrics: type=GAUGE, name=application_1491208276448_0097.driver.jvm.total.used, value=279069288
17/04/04 10:47:36 INFO metrics: type=HISTOGRAM, name=application_1491208276448_0097.driver.CodeGenerator.compilationTime, count=1, min=148, max=148, mean=148.0, stddev=0.0, median=148.0, p75=148.0, p95=148.0, p98=148.0, p99=148.0, p999=148.0
17/04/04 10:47:36 INFO metrics: type=HISTOGRAM, name=application_1491208276448_0097.driver.CodeGenerator.generatedClassSize, count=2, min=532, max=2376, mean=1454.0, stddev=922.0, median=2376.0, p75=2376.0, p95=2376.0, p98=2376.0, p99=2376.0, p999=2376.0
17/04/04 10:47:36 INFO metrics: type=HISTOGRAM, name=application_1491208276448_0097.driver.CodeGenerator.generatedMethodSize, count=5, min=5, max=221, mean=64.2, stddev=81.832511876393, median=15.0, p75=70.0, p95=221.0, p98=221.0, p99=221.0, p999=221.0
17/04/04 10:47:36 INFO metrics: type=HISTOGRAM, name=application_1491208276448_0097.driver.CodeGenerator.sourceCodeSize, count=1, min=1962, max=1962, mean=1962.0, stddev=0.0, median=1962.0, p75=1962.0, p95=1962.0, p98=1962.0, p99=1962.0, p999=1962.0
17/04/04 10:47:36 INFO metrics: type=TIMER, name=application_1491208276448_0097.driver.DAGScheduler.messageProcessingTime, count=20, min=0.0247, max=105.177696, mean=9.992431606157076, stddev=27.565178565314206, median=0.38521, p75=1.296533, p95=79.069126, p98=105.177696, p99=105.177696, p999=105.177696, mean_rate=0.4768138149475206, m1=0.21203362042935395, m5=0.045742426543696334, m15=0.01545137723000816, rate_unit=events/second, duration_unit=milliseconds
17/04/04 10:47:36 INFO TaskSetManager: Lost task 1.1 in stage 0.0 (TID 6) on executor 10.0.0.5: org.apache.spark.SparkException (Task failed while writing rows) [duplicate 5]
17/04/04 10:47:36 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerTaskEnd(0,0,ResultTask,ExceptionFailure(org.apache.spark.SparkException,Task failed while writing rows,[Ljava.lang.StackTraceElement;@6e7a81f0,org.apache.spark.SparkException: Task failed while writing rows
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:261)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(InsertIntoHadoopFsRelationCommand.scala:143)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.spark.TaskKilledException
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply$mcV$sp(WriterContainer.scala:253)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer$$anonfun$writeRows$1.apply(WriterContainer.scala:252)
at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1348)
at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:258)
... 8 more
,Some(org.apache.spark.ThrowableSerializationWrapper@73d4f240),Vector(AccumulableInfo(3,Some(internal.metrics.executorRunTime),Some(2798),None,true,true,None), AccumulableInfo(4,Some(internal.metrics.resultSize),Some(0),None,true,true,None), AccumulableInfo(5,Some(internal.metrics.jvmGCTime),Some(37),None,true,true,None), AccumulableInfo(20,Some(internal.metrics.input.bytesRead),Some(52690944),None,true,true,None)),Vector(LongAccumulator(id: 3, name: Some(internal.metrics.executorRunTime), value: 2798), LongAccumulator(id: 4, name: Some(internal.metrics.resultSize), value: 0), LongAccumulator(id: 5, name: Some(internal.metrics.jvmGCTime), value: 37), LongAccumulator(id: 20, name: Some(internal.metrics.input.bytesRead), value: 52690944))),org.apache.spark.scheduler.TaskInfo@7ce29ac9,org.apache.spark.executor.TaskMetrics@45ffb6b9)
17/04/04 10:47:36 INFO YarnClientSchedulerBackend: Interrupting monitor thread
17/04/04 10:47:36 INFO YarnClientSchedulerBackend: Shutting down all executors
17/04/04 10:47:36 INFO YarnSchedulerBackend$YarnDriverEndpoint: Asking each executor to shut down
17/04/04 10:47:36 INFO SchedulerExtensionServices: Stopping SchedulerExtensionServices
(serviceOption=None,
services=List(),
started=false)
17/04/04 10:47:36 INFO YarnClientSchedulerBackend: Stopped
17/04/04 10:47:36 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
17/04/04 10:47:36 INFO MemoryStore: MemoryStore cleared
17/04/04 10:47:36 INFO BlockManager: BlockManager stopped
17/04/04 10:47:36 INFO BlockManagerMaster: BlockManagerMaster stopped
17/04/04 10:47:36 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
17/04/04 10:47:36 INFO SparkContext: Successfully stopped SparkContext
17/04/04 10:47:36 INFO ShutdownHookManager: Shutdown hook called
17/04/04 10:47:36 INFO ShutdownHookManager: Deleting directory /tmp/spark-42ea5e7c-0997-4a1e-9ff2-881cff8a04d2
17/04/04 10:47:36 INFO MetricsSystemImpl: Stopping azure-file-system metrics system...
17/04/04 10:47:36 INFO MetricsSinkAdapter: azurefs2 thread interrupted.
17/04/04 10:47:36 INFO MetricsSystemImpl: azure-file-system metrics system stopped.
17/04/04 10:47:36 INFO MetricsSystemImpl: azure-file-system metr