Ashish,

I was continuously collecting the details of heap memory "jstat -gc
".During flume stops processing the event processing  i don't see any peak
in heap usage.just 66% heap ratio utilization.

Now I have enabled GC logging with -verbose:gc -XX:+PrintGCDetails
-XX:+PrintGCTimeStamps -Xloggc:<log_file>

let me know if i have any other option for getting better idea why jvm
stops in mid.


On Wed, Jul 30, 2014 at 9:36 PM, SaravanaKumar TR <saran0081...@gmail.com>
wrote:

> It looks like JvisualVM is a GUI tool.
>
> can you please suggest me some CLI tool to get more info about JVM process.
> It looks JVM was hanged because jstack <pid> doesn't show any ..only -F
> option could get some output.
>
>
> On Wed, Jul 30, 2014 at 7:45 PM, Ashish <paliwalash...@gmail.com> wrote:
>
>> Watch for GC activity (either enable GC logging or use visualvm)
>> VisualVM would give complete idea of what's happening inside the running
>> JVM
>>
>>
>> On Tue, Jul 29, 2014 at 11:53 AM, SaravanaKumar TR <
>> saran0081...@gmail.com> wrote:
>>
>>> Out of curiosity to find the exact issue , i started flume with 20 MB
>>> memory (default) with the option as /cv/jvendor/bin/java -Xms20m -Xmx20m
>>> -Dcom.sun.management.jmxremote -XX:-HeapDumpOnOutOfMemoryError
>>> -XX:HeapDumpPath=./java_pid<pid>.hprof
>>>
>>> Again memory channel stops processing the event after a while , but heap
>>> dump is  not generated here.Is the arguments i provided is  corrrect ?
>>>
>>> Tried getting the jstack of running flume agent and seems the process
>>> got hung."tail -F <logfile> process is no more running.
>>>
>>> From jstack -F <pid> ,
>>>
>>> Deadlock Detection:
>>>
>>> No deadlocks found.
>>>
>>> Thread 14788: (state = BLOCKED)
>>>  - sun.misc.Unsafe.park(boolean, long) @bci=0 (Compiled frame;
>>> information may be imprecise)
>>>  - java.util.concurrent.locks.LockSupport.park(java.lang.Object)
>>> @bci=14, line=186 (Compiled frame)
>>>  -
>>> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await()
>>> @bci=42, line=2043 (Compiled frame)
>>>  -
>>> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take()
>>> @bci=98, line=1085 (Compiled frame)
>>>  -
>>> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take()
>>> @bci=1, line=807 (Interpreted frame)
>>>  - java.util.concurrent.ThreadPoolExecutor.getTask() @bci=156, line=1068
>>> (Interpreted frame)
>>>  -
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(java.util.concurrent.ThreadPoolExecutor$Worker)
>>> @bci=26, line=1130 (Interpreted frame)
>>>  - java.util.concurrent.ThreadPoolExecutor$Worker.run() @bci=5, line=615
>>> (Interpreted frame)
>>>  - java.lang.Thread.run() @bci=11, line=744 (Interpreted frame)
>>>
>>>
>>> Thread 14785: (state = IN_NATIVE)
>>>  - java.net.PlainSocketImpl.socketAccept(java.net.SocketImpl) @bci=0
>>> (Interpreted frame)
>>>  - java.net.AbstractPlainSocketImpl.accept(java.net.SocketImpl) @bci=7,
>>> line=398 (Interpreted frame)
>>>  - java.net.ServerSocket.implAccept(java.net.Socket) @bci=60, line=530
>>> (Interpreted frame)
>>>  - java.net.ServerSocket.accept() @bci=48, line=498 (Interpreted frame)
>>>  - sun.management.jmxremote.LocalRMIServerSocketFactory$1.accept()
>>> @bci=1, line=52 (Interpreted frame)
>>>  - sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop()
>>> @bci=55, line=388 (Interpreted frame)
>>>  - sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run() @bci=1, line=360
>>> (Interpreted frame)
>>>  - java.lang.Thread.run() @bci=11, line=744 (Interpreted frame)
>>>
>>>
>>>
>>> On Thu, Jul 17, 2014 at 8:21 PM, SaravanaKumar TR <
>>> saran0081...@gmail.com> wrote:
>>>
>>>> Thanks Natty & Ashish.
>>>>
>>>> I have restarted flume agent with below config.Will monitor it for
>>>> couple of days whether it stops randomly.
>>>>
>>>> JAVA_OPTS="-Xms1g -Xmx1g -Dcom.sun.management.jmxremote
>>>> -XX:-HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./java_pid<pid>.hprof"
>>>>
>>>> thanks for all again.Hope this will work well.
>>>>
>>>>
>>>> On 17 July 2014 12:24, SaravanaKumar TR <saran0081...@gmail.com> wrote:
>>>>
>>>>> thanks its really helpful.
>>>>> I guess the default heap dump path is /tmp ?
>>>>>
>>>>>
>>>>> On 17 July 2014 12:11, Ashish <paliwalash...@gmail.com> wrote:
>>>>>
>>>>>> Nope, a heap dump shall be generated. Please see more options at
>>>>>> http://www.oracle.com/technetwork/java/javase/tech/vmoptions-jsp-140102.html
>>>>>>
>>>>>> to specify path use this -XX:HeapDumpPath=./java_pid<pid>.hprof
>>>>>>
>>>>>>
>>>>>> On Thu, Jul 17, 2014 at 12:09 PM, SaravanaKumar TR <
>>>>>> saran0081...@gmail.com> wrote:
>>>>>>
>>>>>>> yes , sorry I missed to update as 1 GB.
>>>>>>>
>>>>>>> But for out of memory error ,do we get notified in flume logs? I
>>>>>>> haven't see any exception till now.
>>>>>>>
>>>>>>>
>>>>>>> On 17 July 2014 11:55, SaravanaKumar TR <saran0081...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Thanks Ashish , So I wil go ahead and update the flume-env,sh file
>>>>>>>> with
>>>>>>>>
>>>>>>>> JAVA_OPTS="-Xms100m -Xmx200m -Dcom.sun.management.jmxremote
>>>>>>>> -XX:-HeapDumpOnOutOfMemoryError"
>>>>>>>>
>>>>>>>>
>>>>>>>> On 17 July 2014 11:39, Ashish <paliwalash...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Add -XX:-HeapDumpOnOutOfMemoryError parameter as well, if your
>>>>>>>>> process is OOME, would generate a Heap dump. Allocate Heap based on 
>>>>>>>>> the
>>>>>>>>> number of events you need to keep in channel. Try with 1 GB, but 
>>>>>>>>> calculate
>>>>>>>>> according the Channel size as (average event size * number of 
>>>>>>>>> events), plus
>>>>>>>>> object over heads.
>>>>>>>>>
>>>>>>>>> Please note, this is just a rough calculation, actual memory usage
>>>>>>>>> would be higher.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Thu, Jul 17, 2014 at 11:21 AM, SaravanaKumar TR <
>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Okay thanks , So for 128 GB , I will allocate 1 GB as a heap
>>>>>>>>>> memory for flume agent.
>>>>>>>>>>
>>>>>>>>>> But I am surprised why there was no error registered for this
>>>>>>>>>> memory issues in log file (flume.log).
>>>>>>>>>>
>>>>>>>>>> Do i need to check in any other logs?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On 16 July 2014 21:55, Jonathan Natkins <na...@streamsets.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> That's definitely your problem. 20MB is way too low for this.
>>>>>>>>>>> Depending on the other processes you're running with your system, 
>>>>>>>>>>> the
>>>>>>>>>>> amount of memory you'll need will vary, but I'd recommend at least 
>>>>>>>>>>> 1GB. You
>>>>>>>>>>> should define it exactly where it's defined right now, so instead 
>>>>>>>>>>> of the
>>>>>>>>>>> current command, you can run:
>>>>>>>>>>>
>>>>>>>>>>> "/cv/jvendor/bin/java -Xmx1g
>>>>>>>>>>> -Dflume.root.logger=DEBUG,LOGFILE......"
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Jul 16, 2014 at 3:03 AM, SaravanaKumar TR <
>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> I guess i am using defaulk values , from running flume i could
>>>>>>>>>>>> see these lines  "/cv/jvendor/bin/java -Xmx20m
>>>>>>>>>>>> -Dflume.root.logger=DEBUG,LOGFILE......"
>>>>>>>>>>>>
>>>>>>>>>>>> so i guess it takes 20 mb as agent flume memory.
>>>>>>>>>>>> My RAM is 128 GB.So please suggest how much can i assign as
>>>>>>>>>>>> heap memory and where to define it.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 16 July 2014 15:05, Jonathan Natkins <na...@streamsets.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hey Saravana,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I'm attempting to reproduce this, but do you happen to know
>>>>>>>>>>>>> what the Java heap size is for your Flume agent? This information 
>>>>>>>>>>>>> leads me
>>>>>>>>>>>>> to believe that you don't have enough memory allocated to the 
>>>>>>>>>>>>> agent, which
>>>>>>>>>>>>> you may need to do with the -Xmx parameter when you start up your 
>>>>>>>>>>>>> agent.
>>>>>>>>>>>>> That aside, you can set the byteCapacity parameter on the memory 
>>>>>>>>>>>>> channel to
>>>>>>>>>>>>> specify how much memory it is allowed to use. It should default 
>>>>>>>>>>>>> to 80% of
>>>>>>>>>>>>> the Java heap size, but if your heap is too small, this might be 
>>>>>>>>>>>>> a cause of
>>>>>>>>>>>>> errors.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Does anything get written to the log when you try to pass in
>>>>>>>>>>>>> an event of this size?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Natty
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Wed, Jul 16, 2014 at 1:46 AM, SaravanaKumar TR <
>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Natty,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> While looking further , i could see memory channal stops if a
>>>>>>>>>>>>>> line comes with greater than 2 MB.Let me know which parameter 
>>>>>>>>>>>>>> helps us to
>>>>>>>>>>>>>> define max event size of about 3 MB.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 16 July 2014 12:46, SaravanaKumar TR <
>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I am asking point 1 , because in some cases  I could see a
>>>>>>>>>>>>>>> line in logfile around 2 MB.So i need to know what mamimum 
>>>>>>>>>>>>>>> event size.How
>>>>>>>>>>>>>>> to measure it?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 16 July 2014 10:18, SaravanaKumar TR <
>>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi Natty,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Please help me to get the answers for the below queries.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 1,In case of exec source , (tail -F <logfile>) , is that
>>>>>>>>>>>>>>>> each line in file is considered to be a single event ?
>>>>>>>>>>>>>>>> If suppose a line is considered to be a event , what is
>>>>>>>>>>>>>>>> that maximum size of event supported by flume?I mean maximum 
>>>>>>>>>>>>>>>> characters in
>>>>>>>>>>>>>>>> a line supported?
>>>>>>>>>>>>>>>> 2.When event stop processing , I am not seeing "tail -F"
>>>>>>>>>>>>>>>> command running in the background.
>>>>>>>>>>>>>>>> I have used option like "a1.sources.r1.restart = true
>>>>>>>>>>>>>>>> a1.sources.r1.logStdErr = true"..
>>>>>>>>>>>>>>>> Does these config will not send any errors to flume.log if
>>>>>>>>>>>>>>>> any issues in tail?
>>>>>>>>>>>>>>>> Will this config doesnt try to restart the "tail -F" if its
>>>>>>>>>>>>>>>> not running in the background.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 3.Does flume supports all formats of data in logfile or it
>>>>>>>>>>>>>>>> has any predefined data formats..
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Please help me with these to understand better..
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On 16 July 2014 00:56, Jonathan Natkins <
>>>>>>>>>>>>>>>> na...@streamsets.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Saravana,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Everything here looks pretty sane. Do you have a record of
>>>>>>>>>>>>>>>>> the events that came in leading up to the agent stopping 
>>>>>>>>>>>>>>>>> collection? If you
>>>>>>>>>>>>>>>>> can provide the last file created by the agent, and ideally 
>>>>>>>>>>>>>>>>> whatever events
>>>>>>>>>>>>>>>>> had come in, but not been written out to your HDFS sink, it 
>>>>>>>>>>>>>>>>> might be
>>>>>>>>>>>>>>>>> possible for me to reproduce this issue. Would it be possible 
>>>>>>>>>>>>>>>>> to get some
>>>>>>>>>>>>>>>>> sample data from you?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>> Natty
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Tue, Jul 15, 2014 at 10:26 AM, SaravanaKumar TR <
>>>>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi Natty ,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Just to understand , at present my settings is as
>>>>>>>>>>>>>>>>>> "flume.root.logger=INFO,LOGFILE"
>>>>>>>>>>>>>>>>>> in log4j.properties , do you want me to change it to
>>>>>>>>>>>>>>>>>> "flume.root.logger=DEBUG,LOGFILE" and restart the agent.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> But when I start agent , I am already starting with below
>>>>>>>>>>>>>>>>>> command.I guess i am using DEBUG already but not in config 
>>>>>>>>>>>>>>>>>> file , while
>>>>>>>>>>>>>>>>>> starting agent.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ../bin/flume-ng agent -c /d0/flume/conf -f
>>>>>>>>>>>>>>>>>> /d0/flume/conf/flume-conf.properties -n a1 
>>>>>>>>>>>>>>>>>> -Dflume.root.logger=DEBUG,LOGFILE
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> If I do some changes in config "flume-conf.properties" or
>>>>>>>>>>>>>>>>>> restart the agent , it works again and starts collecting the 
>>>>>>>>>>>>>>>>>> data.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> currently all my logs move to flume.log , I dont see any
>>>>>>>>>>>>>>>>>> exception .
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> cat flume.log | grep "Exception"  doesnt show any.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On 15 July 2014 22:24, Jonathan Natkins <
>>>>>>>>>>>>>>>>>> na...@streamsets.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Hi Saravana,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Our best bet on figuring out what's going on here may be
>>>>>>>>>>>>>>>>>>> to turn on the debug logging. What I would recommend is 
>>>>>>>>>>>>>>>>>>> stopping your
>>>>>>>>>>>>>>>>>>> agents, and modifying the log4j properties to turn on DEBUG 
>>>>>>>>>>>>>>>>>>> logging for the
>>>>>>>>>>>>>>>>>>> root logger, and then restart the agents. Once the agent 
>>>>>>>>>>>>>>>>>>> stops producing
>>>>>>>>>>>>>>>>>>> new events, send out the logs and I'll be happy to take a 
>>>>>>>>>>>>>>>>>>> look over them.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Does the system begin working again if you restart the
>>>>>>>>>>>>>>>>>>> agents? Have you noticed any other events correlated with 
>>>>>>>>>>>>>>>>>>> the agent
>>>>>>>>>>>>>>>>>>> stopping collecting events? Maybe a spike in events or 
>>>>>>>>>>>>>>>>>>> something like that?
>>>>>>>>>>>>>>>>>>> And for my own peace of mind, if you run `cat 
>>>>>>>>>>>>>>>>>>> /var/log/flume-ng/* | grep
>>>>>>>>>>>>>>>>>>> "Exception"`, does it bring anything back?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>>>> Natty
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Tue, Jul 15, 2014 at 2:55 AM, SaravanaKumar TR <
>>>>>>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Hi Natty,
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> This is my entire config file.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> # Name the components on this agent
>>>>>>>>>>>>>>>>>>>> a1.sources = r1
>>>>>>>>>>>>>>>>>>>> a1.sinks = k1
>>>>>>>>>>>>>>>>>>>> a1.channels = c1
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> # Describe/configure the source
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.type = exec
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.command = tail -F /data/logs/test_log
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.restart = true
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.logStdErr = true
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> #a1.sources.r1.batchSize = 2
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.interceptors = i1
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.interceptors.i1.type = regex_filter
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.interceptors.i1.regex = resuming normal
>>>>>>>>>>>>>>>>>>>> operations|Received|Response
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> #a1.sources.r1.interceptors = i2
>>>>>>>>>>>>>>>>>>>> #a1.sources.r1.interceptors.i2.type = timestamp
>>>>>>>>>>>>>>>>>>>> #a1.sources.r1.interceptors.i2.preserveExisting = true
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> # Describe the sink
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.type = hdfs
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.path = hdfs://
>>>>>>>>>>>>>>>>>>>> testing.sck.com:9000/running/test.sck/date=%Y-%m-%d
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.writeFormat = Text
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.fileType = DataStream
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.filePrefix = events-
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.rollInterval = 600
>>>>>>>>>>>>>>>>>>>> ##need to run hive query randomly to check teh long
>>>>>>>>>>>>>>>>>>>> running process , so we  need to commit events in hdfs 
>>>>>>>>>>>>>>>>>>>> files regularly
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.rollCount = 0
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.batchSize = 10
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.rollSize = 0
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.hdfs.useLocalTimeStamp = true
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> # Use a channel which buffers events in memory
>>>>>>>>>>>>>>>>>>>> a1.channels.c1.type = memory
>>>>>>>>>>>>>>>>>>>> a1.channels.c1.capacity = 10000
>>>>>>>>>>>>>>>>>>>> a1.channels.c1.transactionCapacity = 10000
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> # Bind the source and sink to the channel
>>>>>>>>>>>>>>>>>>>> a1.sources.r1.channels = c1
>>>>>>>>>>>>>>>>>>>> a1.sinks.k1.channel = c1
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On 14 July 2014 22:54, Jonathan Natkins <
>>>>>>>>>>>>>>>>>>>> na...@streamsets.com> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Hi Saravana,
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> What does your sink configuration look like?
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>> Natty
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On Fri, Jul 11, 2014 at 11:05 PM, SaravanaKumar TR <
>>>>>>>>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Assuming each line in the logfile is considered as a
>>>>>>>>>>>>>>>>>>>>>> event for flume ,
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> 1.Do we have any maximum size of event defined for
>>>>>>>>>>>>>>>>>>>>>> memory/file channel.like any maximum no of characters in 
>>>>>>>>>>>>>>>>>>>>>> a line.
>>>>>>>>>>>>>>>>>>>>>> 2.Does flume supports all formats of data to be
>>>>>>>>>>>>>>>>>>>>>> processed as events or do we have any limitation.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> I am just still trying to understanding why the flume
>>>>>>>>>>>>>>>>>>>>>> stops processing events after sometime.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Can someone please help me out here.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>>> saravana
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> On 11 July 2014 17:49, SaravanaKumar TR <
>>>>>>>>>>>>>>>>>>>>>> saran0081...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Hi ,
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> I am new to flume and  using Apache Flume 1.5.0.
>>>>>>>>>>>>>>>>>>>>>>> Quick setup explanation here.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Source:exec , tail –F command for a logfile.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Channel: tried with both Memory & file channel
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Sink: HDFS
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> When flume starts , processing events happens
>>>>>>>>>>>>>>>>>>>>>>> properly and its moved to hdfs without any issues.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> But after sometime flume suddenly stops sending
>>>>>>>>>>>>>>>>>>>>>>> events to HDFS.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> I am not seeing any errors in logfile flume.log as
>>>>>>>>>>>>>>>>>>>>>>> well.Please let me know if I am missing any 
>>>>>>>>>>>>>>>>>>>>>>> configuration here.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Below is the channel configuration defined and I
>>>>>>>>>>>>>>>>>>>>>>> left the remaining to be default values.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> a1.channels.c1.type = FILE
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> a1.channels.c1.transactionCapacity = 100000
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> a1.channels.c1.capacity = 10000000
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>>>> Saravana
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> thanks
>>>>>>>>> ashish
>>>>>>>>>
>>>>>>>>> Blog: http://www.ashishpaliwal.com/blog
>>>>>>>>> My Photo Galleries: http://www.pbase.com/ashishpaliwal
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> thanks
>>>>>> ashish
>>>>>>
>>>>>> Blog: http://www.ashishpaliwal.com/blog
>>>>>> My Photo Galleries: http://www.pbase.com/ashishpaliwal
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>>
>> --
>> thanks
>> ashish
>>
>> Blog: http://www.ashishpaliwal.com/blog
>> My Photo Galleries: http://www.pbase.com/ashishpaliwal
>>
>
>

Reply via email to