Thanks On Fri, 18 Dec 2020, 00:30 Patrick McCarthy, <pmccar...@dstillery.com> wrote:
> Possibly. In that case maybe you should step back from spark and see if > there are OS-level tools to understand what's going on, like looking for > evidence of the OOM killer - > https://docs.memset.com/other/linux-s-oom-process-killer > > On Thu, Dec 17, 2020 at 1:45 PM Vikas Garg <sperry...@gmail.com> wrote: > >> I am running code in a local machine that is single node machine. >> >> Getting into logs, it looked like the host is killed. This is happening >> very frequently an I am unable to find the reason of this. >> >> Could low memory be the reason? >> >> On Fri, 18 Dec 2020, 00:11 Patrick McCarthy, <pmccar...@dstillery.com> >> wrote: >> >>> 'Job aborted due to stage failure: Task 1 in stage 39.0 failed 1 times' >>> >>> You may want to change the number of failures to a higher number like 4. >>> A single failure on a task should be able to be tolerated, especially if >>> you're on a shared cluster where resources can be preempted. >>> >>> It seems that a node dies or goes off the network, so perhaps you can >>> also debug the logs on the failing node to see why it disappears and >>> prevent the failures in the first place. >>> >>> On Thu, Dec 17, 2020 at 1:27 PM Vikas Garg <sperry...@gmail.com> wrote: >>> >>>> Mydomain is named by me while pasting the logs >>>> >>>> Also, there are multiple class files in my project, if I run any 1 or >>>> 2 at a time, then they run fine, sometimes they too give this error. But >>>> running all the classes at the same time always give this error. >>>> >>>> Once this error come, I can't run any program and on restarting the >>>> system, program starts running fine. >>>> This error goes away on >>>> >>>> On Thu, 17 Dec 2020, 23:50 Patrick McCarthy, <pmccar...@dstillery.com> >>>> wrote: >>>> >>>>> my-domain.com/192.168.166.8:63534 probably isn't a valid address on >>>>> your network, is it? >>>>> >>>>> On Thu, Dec 17, 2020 at 3:03 AM Vikas Garg <sperry...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> Since last few days, I am getting error message while running my >>>>>> project. I have searched Google for the solution but didn't got any help. >>>>>> >>>>>> Can someone help me to figure out how I could mitigate this issue? >>>>>> >>>>>> >>>>>> 20/12/17 13:26:57 ERROR RetryingBlockFetcher: Exception while >>>>>> beginning fetch of 1 outstanding blocks >>>>>> >>>>>> *java.io.IOException*: Failed to connect to >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:253*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:195*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>>>> *NettyBlockTransferService.scala:122*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>>>> *RetryingBlockFetcher.java:141*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.start( >>>>>> *RetryingBlockFetcher.java:121*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.netty.NettyBlockTransferService.fetchBlocks( >>>>>> *NettyBlockTransferService.scala:143*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.BlockTransferService.fetchBlockSync( >>>>>> *BlockTransferService.scala:103*) >>>>>> >>>>>> at >>>>>> org.apache.spark.storage.BlockManager.fetchRemoteManagedBuffer( >>>>>> *BlockManager.scala:1010*) >>>>>> >>>>>> at >>>>>> org.apache.spark.storage.BlockManager.$anonfun$getRemoteBlock$8( >>>>>> *BlockManager.scala:954*) >>>>>> >>>>>> at scala.Option.orElse(*Option.scala:447*) >>>>>> >>>>>> at org.apache.spark.storage.BlockManager.getRemoteBlock( >>>>>> *BlockManager.scala:954*) >>>>>> >>>>>> at org.apache.spark.storage.BlockManager.getRemoteBytes( >>>>>> *BlockManager.scala:1092*) >>>>>> >>>>>> at >>>>>> org.apache.spark.scheduler.TaskResultGetter$$anon$3.$anonfun$run$1( >>>>>> *TaskResultGetter.scala:88*) >>>>>> >>>>>> at >>>>>> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) >>>>>> >>>>>> at org.apache.spark.util.Utils$.logUncaughtExceptions( >>>>>> *Utils.scala:1932*) >>>>>> >>>>>> at org.apache.spark.scheduler.TaskResultGetter$$anon$3.run( >>>>>> *TaskResultGetter.scala:63*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>> *ThreadPoolExecutor.java:1149*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>> *ThreadPoolExecutor.java:624*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> Caused by: >>>>>> *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>>>> Permission denied: no further information: >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>>>> information >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>>>> *SocketChannelImpl.java:715*) >>>>>> >>>>>> at >>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>>>> *NioSocketChannel.java:330*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>>>> *AbstractNioChannel.java:334*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>>>> *NioEventLoop.java:702*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>>>> *NioEventLoop.java:650*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>>>> *NioEventLoop.java:576*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.run( >>>>>> *NioEventLoop.java:493*) >>>>>> >>>>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>>>> *SingleThreadEventExecutor.java:989*) >>>>>> >>>>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>>>> *ThreadExecutorMap.java:74*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> 20/12/17 13:27:02 ERROR RetryingBlockFetcher: Exception while >>>>>> beginning fetch of 1 outstanding blocks (after 1 retries) >>>>>> >>>>>> *java.io.IOException*: Failed to connect to >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:253*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:195*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>>>> *NettyBlockTransferService.scala:122*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>>>> *RetryingBlockFetcher.java:141*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>>>>> *RetryingBlockFetcher.java:169*) >>>>>> >>>>>> at java.util.concurrent.Executors$RunnableAdapter.call( >>>>>> *Executors.java:511*) >>>>>> >>>>>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>> *ThreadPoolExecutor.java:1149*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>> *ThreadPoolExecutor.java:624*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> Caused by: >>>>>> *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>>>> Permission denied: no further information: >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>>>> information >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>>>> *SocketChannelImpl.java:715*) >>>>>> >>>>>> at >>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>>>> *NioSocketChannel.java:330*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>>>> *AbstractNioChannel.java:334*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>>>> *NioEventLoop.java:702*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>>>> *NioEventLoop.java:650*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>>>> *NioEventLoop.java:576*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.run( >>>>>> *NioEventLoop.java:493*) >>>>>> >>>>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>>>> *SingleThreadEventExecutor.java:989*) >>>>>> >>>>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>>>> *ThreadExecutorMap.java:74*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> 20/12/17 13:27:07 ERROR RetryingBlockFetcher: Exception while >>>>>> beginning fetch of 1 outstanding blocks (after 2 retries) >>>>>> >>>>>> *java.io.IOException*: Failed to connect to >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:253*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:195*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>>>> *NettyBlockTransferService.scala:122*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>>>> *RetryingBlockFetcher.java:141*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>>>>> *RetryingBlockFetcher.java:169*) >>>>>> >>>>>> at java.util.concurrent.Executors$RunnableAdapter.call( >>>>>> *Executors.java:511*) >>>>>> >>>>>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>> *ThreadPoolExecutor.java:1149*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>> *ThreadPoolExecutor.java:624*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> Caused by: >>>>>> *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>>>> Permission denied: no further information: >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>>>> information >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>>>> *SocketChannelImpl.java:715*) >>>>>> >>>>>> at >>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>>>> *NioSocketChannel.java:330*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>>>> *AbstractNioChannel.java:334*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>>>> *NioEventLoop.java:702*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>>>> *NioEventLoop.java:650*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>>>> *NioEventLoop.java:576*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.run( >>>>>> *NioEventLoop.java:493*) >>>>>> >>>>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>>>> *SingleThreadEventExecutor.java:989*) >>>>>> >>>>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>>>> *ThreadExecutorMap.java:74*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> 20/12/17 13:27:12 ERROR RetryingBlockFetcher: Exception while >>>>>> beginning fetch of 1 outstanding blocks (after 3 retries) >>>>>> >>>>>> *java.io.IOException*: Failed to connect to >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:253*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.client.TransportClientFactory.createClient( >>>>>> *TransportClientFactory.java:195*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.netty.NettyBlockTransferService$$anon$2.createAndStart( >>>>>> *NettyBlockTransferService.scala:122*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.fetchAllOutstanding( >>>>>> *RetryingBlockFetcher.java:141*) >>>>>> >>>>>> at >>>>>> org.apache.spark.network.shuffle.RetryingBlockFetcher.lambda$initiateRetry$0( >>>>>> *RetryingBlockFetcher.java:169*) >>>>>> >>>>>> at java.util.concurrent.Executors$RunnableAdapter.call( >>>>>> *Executors.java:511*) >>>>>> >>>>>> at java.util.concurrent.FutureTask.run(*FutureTask.java:266*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>> *ThreadPoolExecutor.java:1149*) >>>>>> >>>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>> *ThreadPoolExecutor.java:624*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> Caused by: >>>>>> *io.netty.channel.AbstractChannel$AnnotatedSocketException*: >>>>>> Permission denied: no further information: >>>>>> my-domain.com/192.168.166.8:63534 >>>>>> >>>>>> Caused by: *java.net.SocketException*: Permission denied: no further >>>>>> information >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(*Native Method*) >>>>>> >>>>>> at sun.nio.ch.SocketChannelImpl.finishConnect( >>>>>> *SocketChannelImpl.java:715*) >>>>>> >>>>>> at >>>>>> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect( >>>>>> *NioSocketChannel.java:330*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect( >>>>>> *AbstractNioChannel.java:334*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKey( >>>>>> *NioEventLoop.java:702*) >>>>>> >>>>>> at >>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized( >>>>>> *NioEventLoop.java:650*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.processSelectedKeys( >>>>>> *NioEventLoop.java:576*) >>>>>> >>>>>> at io.netty.channel.nio.NioEventLoop.run( >>>>>> *NioEventLoop.java:493*) >>>>>> >>>>>> at io.netty.util.concurrent.SingleThreadEventExecutor$4.run( >>>>>> *SingleThreadEventExecutor.java:989*) >>>>>> >>>>>> at io.netty.util.internal.ThreadExecutorMap$2.run( >>>>>> *ThreadExecutorMap.java:74*) >>>>>> >>>>>> at io.netty.util.concurrent.FastThreadLocalRunnable.run( >>>>>> *FastThreadLocalRunnable.java:30*) >>>>>> >>>>>> at java.lang.Thread.run(*Thread.java:748*) >>>>>> >>>>>> 20/12/17 13:27:12 ERROR TaskSetManager: Task 1 in stage 39.0 failed 1 >>>>>> times; aborting job >>>>>> >>>>>> 20/12/17 13:27:12 ERROR FileFormatWriter: Aborting job >>>>>> 0db8a16a-e1d3-4091-b725-0a8bdc17b0b1. >>>>>> >>>>>> *java.util.concurrent.ExecutionException*: >>>>>> *org.apache.spark.SparkException*: Job aborted due to stage failure: >>>>>> Task 1 in stage 39.0 failed 1 times, most recent failure: Lost task 1.0 >>>>>> in >>>>>> stage 39.0 (TID 655, my-domain.com, executor driver): TaskResultLost >>>>>> (result lost from block manager) >>>>>> >>>>>> Driver stacktrace: >>>>>> >>>>>> at java.util.concurrent.FutureTask.report( >>>>>> *FutureTask.java:122*) >>>>>> >>>>>> at java.util.concurrent.FutureTask.get(*FutureTask.java:206*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast( >>>>>> *BroadcastExchangeExec.scala:195*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.InputAdapter.doExecuteBroadcast( >>>>>> *WholeStageCodegenExec.scala:515*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1( >>>>>> *SparkPlan.scala:188*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeBroadcast( >>>>>> *SparkPlan.scala:184*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.prepareBroadcast( >>>>>> *BroadcastHashJoinExec.scala:116*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.codegenInner( >>>>>> *BroadcastHashJoinExec.scala:210*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doConsume( >>>>>> *BroadcastHashJoinExec.scala:100*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>>>>> *WholeStageCodegenExec.scala:194*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>>>>> *WholeStageCodegenExec.scala:149*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ProjectExec.consume( >>>>>> *basicPhysicalOperators.scala:41*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ProjectExec.doConsume( >>>>>> *basicPhysicalOperators.scala:71*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>>>>> *WholeStageCodegenExec.scala:194*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>>>>> *WholeStageCodegenExec.scala:149*) >>>>>> >>>>>> at org.apache.spark.sql.execution.FilterExec.consume( >>>>>> *basicPhysicalOperators.scala:97*) >>>>>> >>>>>> at org.apache.spark.sql.execution.FilterExec.doConsume( >>>>>> *basicPhysicalOperators.scala:222*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.consume( >>>>>> *WholeStageCodegenExec.scala:194*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.consume$( >>>>>> *WholeStageCodegenExec.scala:149*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ColumnarToRowExec.consume( >>>>>> *Columnar.scala:60*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ColumnarToRowExec.doProduce( >>>>>> *Columnar.scala:185*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>>>> *WholeStageCodegenExec.scala:95*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ColumnarToRowExec.produce( >>>>>> *Columnar.scala:60*) >>>>>> >>>>>> at org.apache.spark.sql.execution.FilterExec.doProduce( >>>>>> *basicPhysicalOperators.scala:137*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>>>> *WholeStageCodegenExec.scala:95*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.FilterExec.produce( >>>>>> *basicPhysicalOperators.scala:97*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ProjectExec.doProduce( >>>>>> *basicPhysicalOperators.scala:51*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>>>> *WholeStageCodegenExec.scala:95*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ProjectExec.produce( >>>>>> *basicPhysicalOperators.scala:41*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce( >>>>>> *BroadcastHashJoinExec.scala:95*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>>>> *WholeStageCodegenExec.scala:95*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce( >>>>>> *BroadcastHashJoinExec.scala:39*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ProjectExec.doProduce( >>>>>> *basicPhysicalOperators.scala:51*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.CodegenSupport.$anonfun$produce$1( >>>>>> *WholeStageCodegenExec.scala:95*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.CodegenSupport.produce$( >>>>>> *WholeStageCodegenExec.scala:90*) >>>>>> >>>>>> at org.apache.spark.sql.execution.ProjectExec.produce( >>>>>> *basicPhysicalOperators.scala:41*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen( >>>>>> *WholeStageCodegenExec.scala:632*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute( >>>>>> *WholeStageCodegenExec.scala:692*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1( >>>>>> *SparkPlan.scala:175*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1( >>>>>> *SparkPlan.scala:213*) >>>>>> >>>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope( >>>>>> *RDDOperationScope.scala:151*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.executeQuery( >>>>>> *SparkPlan.scala:210*) >>>>>> >>>>>> at org.apache.spark.sql.execution.SparkPlan.execute( >>>>>> *SparkPlan.scala:171*) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.datasources.FileFormatWriter$.write( >>>>>> *FileFormatWriter.scala:172*) >>>>>> >>>>>>