Thanks Ashish , So I wil go ahead and update the flume-env,sh file with

JAVA_OPTS="-Xms100m -Xmx200m -Dcom.sun.management.jmxremote
-XX:-HeapDumpOnOutOfMemoryError"


On 17 July 2014 11:39, Ashish <paliwalash...@gmail.com> wrote:

> Add -XX:-HeapDumpOnOutOfMemoryError parameter as well, if your process is
> OOME, would generate a Heap dump. Allocate Heap based on the number of
> events you need to keep in channel. Try with 1 GB, but calculate according
> the Channel size as (average event size * number of events), plus object
> over heads.
>
> Please note, this is just a rough calculation, actual memory usage would
> be higher.
>
>
> On Thu, Jul 17, 2014 at 11:21 AM, SaravanaKumar TR <saran0081...@gmail.com
> > wrote:
>
>> Okay thanks , So for 128 GB , I will allocate 1 GB as a heap memory for
>> flume agent.
>>
>> But I am surprised why there was no error registered for this memory
>> issues in log file (flume.log).
>>
>> Do i need to check in any other logs?
>>
>>
>> On 16 July 2014 21:55, Jonathan Natkins <na...@streamsets.com> wrote:
>>
>>> That's definitely your problem. 20MB is way too low for this. Depending
>>> on the other processes you're running with your system, the amount of
>>> memory you'll need will vary, but I'd recommend at least 1GB. You should
>>> define it exactly where it's defined right now, so instead of the current
>>> command, you can run:
>>>
>>> "/cv/jvendor/bin/java -Xmx1g -Dflume.root.logger=DEBUG,LOGFILE......"
>>>
>>>
>>> On Wed, Jul 16, 2014 at 3:03 AM, SaravanaKumar TR <
>>> saran0081...@gmail.com> wrote:
>>>
>>>> I guess i am using defaulk values , from running flume i could see
>>>> these lines  "/cv/jvendor/bin/java -Xmx20m
>>>> -Dflume.root.logger=DEBUG,LOGFILE......"
>>>>
>>>> so i guess it takes 20 mb as agent flume memory.
>>>> My RAM is 128 GB.So please suggest how much can i assign as heap memory
>>>> and where to define it.
>>>>
>>>>
>>>> On 16 July 2014 15:05, Jonathan Natkins <na...@streamsets.com> wrote:
>>>>
>>>>> Hey Saravana,
>>>>>
>>>>> I'm attempting to reproduce this, but do you happen to know what the
>>>>> Java heap size is for your Flume agent? This information leads me to
>>>>> believe that you don't have enough memory allocated to the agent, which 
>>>>> you
>>>>> may need to do with the -Xmx parameter when you start up your agent. That
>>>>> aside, you can set the byteCapacity parameter on the memory channel to
>>>>> specify how much memory it is allowed to use. It should default to 80% of
>>>>> the Java heap size, but if your heap is too small, this might be a cause 
>>>>> of
>>>>> errors.
>>>>>
>>>>> Does anything get written to the log when you try to pass in an event
>>>>> of this size?
>>>>>
>>>>> Thanks,
>>>>> Natty
>>>>>
>>>>>
>>>>> On Wed, Jul 16, 2014 at 1:46 AM, SaravanaKumar TR <
>>>>> saran0081...@gmail.com> wrote:
>>>>>
>>>>>> Hi Natty,
>>>>>>
>>>>>> While looking further , i could see memory channal stops if a line
>>>>>> comes with greater than 2 MB.Let me know which parameter helps us to 
>>>>>> define
>>>>>> max event size of about 3 MB.
>>>>>>
>>>>>>
>>>>>> On 16 July 2014 12:46, SaravanaKumar TR <saran0081...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I am asking point 1 , because in some cases  I could see a line in
>>>>>>> logfile around 2 MB.So i need to know what mamimum event size.How to
>>>>>>> measure it?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On 16 July 2014 10:18, SaravanaKumar TR <saran0081...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi Natty,
>>>>>>>>
>>>>>>>> Please help me to get the answers for the below queries.
>>>>>>>>
>>>>>>>> 1,In case of exec source , (tail -F <logfile>) , is that each line
>>>>>>>> in file is considered to be a single event ?
>>>>>>>> If suppose a line is considered to be a event , what is that
>>>>>>>> maximum size of event supported by flume?I mean maximum characters in a
>>>>>>>> line supported?
>>>>>>>> 2.When event stop processing , I am not seeing "tail -F" command
>>>>>>>> running in the background.
>>>>>>>> I have used option like "a1.sources.r1.restart = true
>>>>>>>> a1.sources.r1.logStdErr = true"..
>>>>>>>> Does these config will not send any errors to flume.log if any
>>>>>>>> issues in tail?
>>>>>>>> Will this config doesnt try to restart the "tail -F" if its not
>>>>>>>> running in the background.
>>>>>>>>
>>>>>>>> 3.Does flume supports all formats of data in logfile or it has any
>>>>>>>> predefined data formats..
>>>>>>>>
>>>>>>>> Please help me with these to understand better..
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On 16 July 2014 00:56, Jonathan Natkins <na...@streamsets.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Saravana,
>>>>>>>>>
>>>>>>>>> Everything here looks pretty sane. Do you have a record of the
>>>>>>>>> events that came in leading up to the agent stopping collection? If 
>>>>>>>>> you can
>>>>>>>>> provide the last file created by the agent, and ideally whatever 
>>>>>>>>> events had
>>>>>>>>> come in, but not been written out to your HDFS sink, it might be 
>>>>>>>>> possible
>>>>>>>>> for me to reproduce this issue. Would it be possible to get some 
>>>>>>>>> sample
>>>>>>>>> data from you?
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Natty
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Jul 15, 2014 at 10:26 AM, SaravanaKumar TR <
>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Natty ,
>>>>>>>>>>
>>>>>>>>>> Just to understand , at present my settings is as
>>>>>>>>>> "flume.root.logger=INFO,LOGFILE"
>>>>>>>>>> in log4j.properties , do you want me to change it to
>>>>>>>>>> "flume.root.logger=DEBUG,LOGFILE" and restart the agent.
>>>>>>>>>>
>>>>>>>>>> But when I start agent , I am already starting with below
>>>>>>>>>> command.I guess i am using DEBUG already but not in config file , 
>>>>>>>>>> while
>>>>>>>>>> starting agent.
>>>>>>>>>>
>>>>>>>>>> ../bin/flume-ng agent -c /d0/flume/conf -f
>>>>>>>>>> /d0/flume/conf/flume-conf.properties -n a1 
>>>>>>>>>> -Dflume.root.logger=DEBUG,LOGFILE
>>>>>>>>>>
>>>>>>>>>> If I do some changes in config "flume-conf.properties" or restart
>>>>>>>>>> the agent , it works again and starts collecting the data.
>>>>>>>>>>
>>>>>>>>>> currently all my logs move to flume.log , I dont see any
>>>>>>>>>> exception .
>>>>>>>>>>
>>>>>>>>>> cat flume.log | grep "Exception"  doesnt show any.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On 15 July 2014 22:24, Jonathan Natkins <na...@streamsets.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Saravana,
>>>>>>>>>>>
>>>>>>>>>>> Our best bet on figuring out what's going on here may be to turn
>>>>>>>>>>> on the debug logging. What I would recommend is stopping your 
>>>>>>>>>>> agents, and
>>>>>>>>>>> modifying the log4j properties to turn on DEBUG logging for the root
>>>>>>>>>>> logger, and then restart the agents. Once the agent stops producing 
>>>>>>>>>>> new
>>>>>>>>>>> events, send out the logs and I'll be happy to take a look over 
>>>>>>>>>>> them.
>>>>>>>>>>>
>>>>>>>>>>> Does the system begin working again if you restart the agents?
>>>>>>>>>>> Have you noticed any other events correlated with the agent stopping
>>>>>>>>>>> collecting events? Maybe a spike in events or something like that? 
>>>>>>>>>>> And for
>>>>>>>>>>> my own peace of mind, if you run `cat /var/log/flume-ng/* | grep
>>>>>>>>>>> "Exception"`, does it bring anything back?
>>>>>>>>>>>
>>>>>>>>>>> Thanks!
>>>>>>>>>>> Natty
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Jul 15, 2014 at 2:55 AM, SaravanaKumar TR <
>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Natty,
>>>>>>>>>>>>
>>>>>>>>>>>> This is my entire config file.
>>>>>>>>>>>>
>>>>>>>>>>>> # Name the components on this agent
>>>>>>>>>>>> a1.sources = r1
>>>>>>>>>>>> a1.sinks = k1
>>>>>>>>>>>> a1.channels = c1
>>>>>>>>>>>>
>>>>>>>>>>>> # Describe/configure the source
>>>>>>>>>>>> a1.sources.r1.type = exec
>>>>>>>>>>>> a1.sources.r1.command = tail -F /data/logs/test_log
>>>>>>>>>>>> a1.sources.r1.restart = true
>>>>>>>>>>>> a1.sources.r1.logStdErr = true
>>>>>>>>>>>>
>>>>>>>>>>>> #a1.sources.r1.batchSize = 2
>>>>>>>>>>>>
>>>>>>>>>>>> a1.sources.r1.interceptors = i1
>>>>>>>>>>>> a1.sources.r1.interceptors.i1.type = regex_filter
>>>>>>>>>>>> a1.sources.r1.interceptors.i1.regex = resuming normal
>>>>>>>>>>>> operations|Received|Response
>>>>>>>>>>>>
>>>>>>>>>>>> #a1.sources.r1.interceptors = i2
>>>>>>>>>>>> #a1.sources.r1.interceptors.i2.type = timestamp
>>>>>>>>>>>> #a1.sources.r1.interceptors.i2.preserveExisting = true
>>>>>>>>>>>>
>>>>>>>>>>>> # Describe the sink
>>>>>>>>>>>> a1.sinks.k1.type = hdfs
>>>>>>>>>>>> a1.sinks.k1.hdfs.path = hdfs://
>>>>>>>>>>>> testing.sck.com:9000/running/test.sck/date=%Y-%m-%d
>>>>>>>>>>>> a1.sinks.k1.hdfs.writeFormat = Text
>>>>>>>>>>>> a1.sinks.k1.hdfs.fileType = DataStream
>>>>>>>>>>>> a1.sinks.k1.hdfs.filePrefix = events-
>>>>>>>>>>>> a1.sinks.k1.hdfs.rollInterval = 600
>>>>>>>>>>>> ##need to run hive query randomly to check teh long running
>>>>>>>>>>>> process , so we  need to commit events in hdfs files regularly
>>>>>>>>>>>> a1.sinks.k1.hdfs.rollCount = 0
>>>>>>>>>>>> a1.sinks.k1.hdfs.batchSize = 10
>>>>>>>>>>>> a1.sinks.k1.hdfs.rollSize = 0
>>>>>>>>>>>> a1.sinks.k1.hdfs.useLocalTimeStamp = true
>>>>>>>>>>>>
>>>>>>>>>>>> # Use a channel which buffers events in memory
>>>>>>>>>>>> a1.channels.c1.type = memory
>>>>>>>>>>>> a1.channels.c1.capacity = 10000
>>>>>>>>>>>> a1.channels.c1.transactionCapacity = 10000
>>>>>>>>>>>>
>>>>>>>>>>>> # Bind the source and sink to the channel
>>>>>>>>>>>> a1.sources.r1.channels = c1
>>>>>>>>>>>> a1.sinks.k1.channel = c1
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 14 July 2014 22:54, Jonathan Natkins <na...@streamsets.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Saravana,
>>>>>>>>>>>>>
>>>>>>>>>>>>> What does your sink configuration look like?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Natty
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Fri, Jul 11, 2014 at 11:05 PM, SaravanaKumar TR <
>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Assuming each line in the logfile is considered as a event
>>>>>>>>>>>>>> for flume ,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 1.Do we have any maximum size of event defined for
>>>>>>>>>>>>>> memory/file channel.like any maximum no of characters in a line.
>>>>>>>>>>>>>> 2.Does flume supports all formats of data to be processed as
>>>>>>>>>>>>>> events or do we have any limitation.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I am just still trying to understanding why the flume stops
>>>>>>>>>>>>>> processing events after sometime.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Can someone please help me out here.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>> saravana
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 11 July 2014 17:49, SaravanaKumar TR <
>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi ,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I am new to flume and  using Apache Flume 1.5.0. Quick setup
>>>>>>>>>>>>>>> explanation here.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Source:exec , tail –F command for a logfile.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Channel: tried with both Memory & file channel
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Sink: HDFS
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> When flume starts , processing events happens properly and
>>>>>>>>>>>>>>> its moved to hdfs without any issues.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> But after sometime flume suddenly stops sending events to
>>>>>>>>>>>>>>> HDFS.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I am not seeing any errors in logfile flume.log as
>>>>>>>>>>>>>>> well.Please let me know if I am missing any configuration here.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Below is the channel configuration defined and I left the
>>>>>>>>>>>>>>> remaining to be default values.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> a1.channels.c1.type = FILE
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> a1.channels.c1.transactionCapacity = 100000
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> a1.channels.c1.capacity = 10000000
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>> Saravana
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
>
> --
> thanks
> ashish
>
> Blog: http://www.ashishpaliwal.com/blog
> My Photo Galleries: http://www.pbase.com/ashishpaliwal
>

Reply via email to