Thank you for replying. 
Actually there is no any message coming during the exception. And there is no 
OOME in any executor. What I am suspecting it might be caused by AWL. 

> On Apr 2, 2017, at 5:22 AM, Timur Shenkao <t...@timshenkao.su> wrote:
> 
> Hello,
> It's difficult to tell without details.
> I believe one of the executors dies because of OOM or some Runtime Exception 
> (some unforeseen dirty data row).
> Less probable is GC stop-the-world pause when incoming message rate increases 
> drastically.
> 
> 
>> On Saturday, April 1, 2017, Jiang Jacky <jiang0...@gmail.com> wrote:
>> Hello, Guys
>> I am running the spark streaming in 2.1.0, the scala version is tried on 
>> 2.11.7 and 2.11.4. And it is consuming from JMS. Recently, I have get the 
>> following error
>> "ERROR scheduler.ReceiverTracker: Deregistered receiver for stream 0: 
>> Stopped by driver"
>> 
>> This error can be occurred randomly, it might be couple hours or couple 
>> days. besides this error, everything is perfect.
>> When the error happens, my job is stopped completely. There is no any other 
>> error can be found.
>> I am running on top of yarn, and tried to look up the error through yarn 
>> logs, container, no any further information appears there. The job is just 
>> stopped from driver gracefully. BTW I have customized receiver, I either do 
>> not think it is happened from receiver, there is no any error exception from 
>> receiver, and I can also track the stop command is sent from "onStop" 
>> function in receiver.
>> 
>> FYI, the driver is not consuming any large memory, there is no any RDD 
>> "collect" command in the driver. I have also checked container log for each 
>> executor, and cannot find any further error.
>> 
>> 
>> 
>> 
>> The following is my conf for the spark context
>> val conf = new SparkConf().setAppName(jobName).setMaster(master)
>>   .set("spark.hadoop.validateOutputSpecs", "false")
>>   .set("spark.driver.allowMultipleContexts", "true")
>>   .set("spark.streaming.receiver.maxRate", "500")
>>   .set("spark.streaming.backpressure.enabled", "true")
>>   .set("spark.streaming.stopGracefullyOnShutdown", "true")
>>   .set("spark.eventLog.enabled", "true");
>> 
>> If you have any idea or suggestion, please let me know. Appreciate on the 
>> solution.
>> 
>> Thank you so much
>> 

Reply via email to