Re: Alerts when Flume agent fails

2017-02-26 Thread Sharninder Khera
Set up scripts to send alerts sooner ? There isn't a built in way in flume so you will have to setup monitoring separately  On Mon, Feb 27, 2017 at 8:57 AM +0530, "Suresh V" wrote: Hello, Is there a way to set up an alert mechanism by email immediately when a flume agent fails d

Re: Flume interseptor and multiplexing

2016-02-16 Thread Sharninder
Not sure I understood the question correctly but you might be able to use the regex interceptor to do what you want. -- Sharninder On Tue, Feb 16, 2016 at 1:56 PM, Gonzalo Herreros wrote: > I don't think that's possible without writing/reusing custom code. > You would need a

Re: How to monitor flume agents

2015-11-19 Thread Sharninder
Monitoring requirements are different for different teams but have you looked at collectd? That with graphite/grafana could work for you. > On 20-Nov-2015, at 10:00 AM, Hemanth Abbina wrote: > > Just a follow up question. > > We need a remote Linux server monitoring tool, to monitor our clu

Re: Batchsize in kafka sink

2015-09-27 Thread Sharninder
ransaction will surely help withe reducing IO on the flume server. Thanks a lot. -- Sharninder On Sun, Sep 27, 2015 at 3:16 PM, Gonzalo Herreros wrote: > There are subtle but significant differences. > > When you configure in the sink: "batchSize" you are specifying how m

Re: Batchsize in kafka sink

2015-09-25 Thread Sharninder
Anyone ? > On 25-Sep-2015, at 3:51 PM, Sharninder wrote: > > Hi, > > We want to move to the built-in kafka sink from our own custom implementation > and I have a question about the batchsize config parameter. > > Looking at the code of the sink, I can tell that

Batchsize in kafka sink

2015-09-25 Thread Sharninder
is the difference between this variable and the kafka batch.num.messages parameter? Is the flume parameter necessary ? -- Sharninder

Re: How to calculate the throughput?

2015-09-23 Thread Sharninder
;, > "StopTime": "0", > "RollbackCount": "0", > "StartTime": "1443053580579", > "BatchUnderflowCount": "0" > }, > "SOURCE.my-kafka-source": { > "KafkaEventGetTimer": "886", > "OpenConnectionCount": "0", > "Type": "SOURCE", > "AppendBatchAcceptedCount": "0", > "AppendBatchReceivedCount": "0", > "EventAcceptedCount": "6500", > "AppendReceivedCount": "0", > "StopTime": "0", > "StartTime": "1443053581472", > "EventReceivedCount": "7000", > "KafkaCommitTimer": "246", > "AppendAcceptedCount": "0" > } > } > > > Thanks! > -- -- Sharninder

Re: Flume Message Headers?

2015-09-19 Thread Sharninder
gt; On Fri, Sep 18, 2015 at 10:28 PM, Sharninder wrote: > >> Use the Kafka console utility to see what's being fed to Kafka. >> >> >> On Saturday, September 19, 2015, Doug McClure wrote: >> >>> Ideally as they appear passing through broker in topic and

Re: Flume Message Headers?

2015-09-18 Thread Sharninder
being collected in the spooldir >> source and sent via kafka sink, specifically the values of fileHeader and >> basenameHeader data. >> >> Tks! >> >> Doug >> > > -- -- Sharninder

Re: Need Urgent Help (please) with HTTP Source/JSON Handling

2015-09-07 Thread Sharninder
implemented and, honestly, it shouldn't be more than a couple of hours of work, most likely less. -- Sharninder On Fri, Sep 4, 2015 at 11:34 PM, Sutanu Das wrote: > Huge THANKS Hari. > > > > I just did this per your recommendation/docs – and it worked !!!, I can > n

Re: A question regarding memory vs file channel in flume

2015-07-29 Thread Sharninder
e generic enough. The file channel is durable but depends on a lot of things including the kind of disks etc. The type of messages being sent might affect the throughput too. In short you'll have to test for your use case and your hardware. -- Sharninder

RE: Flume minimum java heap memory

2015-04-23 Thread Sharninder Khera
The kind of messages and your particular use case will really define the amount of heap you need. Why start with 200m. Set both Xms and Xmx to one value, say 1g and test from there.  Sharninder _ From: mahendran m Sent: Friday, April 24, 2015 9:39 am Subject

Re: flume with gnip

2014-10-17 Thread Sharninder
yes a source. My bad :) On Fri, Oct 17, 2014 at 9:54 PM, Anthony Biacco wrote: > You mean a custom source, right? Not a sink. > > -Tony > > > On Thu, Oct 16, 2014 at 10:47 PM, Sharninder wrote: > >> You need to write a custom sink using the flume sdk. I'm not su

Re: flume with gnip

2014-10-16 Thread Sharninder
You need to write a custom sink using the flume sdk. I'm not sure what are the APIs provided by the gnip4 jar but it shouldn't be too difficult to write the flume part. -- Sharninder On Fri, Oct 17, 2014 at 2:08 AM, Anthony Biacco wrote: > > We were using flume with twitter&#

Re: Flume Syslog source

2014-10-16 Thread Sharninder
Thanks Jeff. I'll take a look at the multipart source too. On Thu, Oct 16, 2014 at 8:52 PM, Jeff Lord wrote: > You will get better perf out of the multiport syslog source > > > On Wednesday, October 15, 2014, Sharninder wrote: > >> I just looked at the existing syslog

Re: Flume Syslog source

2014-10-15 Thread Sharninder
I just looked at the existing syslogtcp source and it seems it does take pains to parse the hostname from the message and I think that is the best bet for me. Ofcourse, it might fail for a few devices, but I'll just have to think of something else for those. -- Sharninder On Thu, Oct 16,

Re: Flume Syslog source

2014-10-15 Thread Sharninder
Yes Jeff. That's a possiblity but I'm not sure (actually pretty sure) that there would be a some random device which will not send their logs in the proper format and my regex will break. This is the way I'll implement it if I can't find anything better. Thanks, Sharninder

Re: Flume Syslog source

2014-10-15 Thread Sharninder
Yes, I did think of that but that just seems like a hack and doesn't scale too much. Ideally, I should be able to just look at the remote host from the tcp headers somewhere and add that info to the flume event header. -- Sharninder On Thu, Oct 16, 2014 at 10:17 AM, Hari Shreed

Flume Syslog source

2014-10-15 Thread Sharninder
eader somehow? I can then use either an interceptor to read the header or even a custom sink to deal with events based on the hostname. -- Sharninder

Re: Better file sink

2014-09-26 Thread Sharninder
Implementing a new sink isn't that difficult really so if your requirements aren't met with the current file sink, by all means write your own. Take a look at the existing code and the Dev guide. > On 26-Sep-2014, at 5:47 pm, jean garutti wrote: > > Hello, > For differents purposes, i want to

Re: flume agent / JVM crashes periodically

2014-09-20 Thread Sharninder
Have you increased the JVM heap size or running with the default 20M only? -- Sharninder > On 21-Sep-2014, at 3:08 am, terreyshih wrote: > > BTW, no exceptions are thrown. It just started the agent-shutdown-hook. > > thx > >> On Sep 20, 2014, at 2:33 PM, terr

Re: Newbie - Sink question

2014-09-04 Thread Sharninder
g else in the interceptor/morphline. -- Sharninder On Fri, Sep 5, 2014 at 8:23 AM, Kevin Warner wrote: > Hello All, > We have the following configuration: > Source->Channel->Sink > > Now, the source is pointing to a folder that has lots of json files. The > channel is file ba

Re: avro source vs syslog source in flume

2014-08-24 Thread Sharninder
Yes, you can compose avro messages in php. As to your question about which is a better performances wise, I'm not really sure about that, but since you already have a script generating syslog, why not just start using it. To add header values to each event you can add an interceptor to the topology

Re: how to load balance flume

2014-08-14 Thread Sharninder
I'm not sure without looking at the exact usecase, but maybe you can use something like haproxy? -- Sharninder On Thu, Aug 14, 2014 at 4:08 PM, Mohit Durgapal wrote: > Hi Sharninder, > > Thanks for the response. The load balancing is not based on header. To > simplify, le

Re: how to load balance flume

2014-08-14 Thread Sharninder
To add headers to the events, you can either send proper avro formatted packets (which have a header) to an avro source, or implement a custom interceptor to add headers after they're received by the syslog source. There is a static interceptor bundled with flume that you can use. The problem with

Re: calling avro RPC in a customized sink

2014-08-12 Thread Sharninder
The purpose of a sink is to take events from a channel and pass it on to a further destination. That destination can be anything, including your another program listening on rpc. On Tue, Aug 12, 2014 at 1:26 PM, terreyshih wrote: > Hi, > > It seems to me that the intention of the flume is to be

Re: Collect TCP data over TCP stream

2014-07-31 Thread Sharninder
Liu, you first need to figure out what TCP data you want to collect. Is there a possibility that this data can be collected at some central router/gateway using SNMP? If not SNMP then you can definitely run something like wireshark or write up your own tool using a library like libpcap and collect

Re: AWS S3 flume source

2014-07-31 Thread Sharninder
As far as I know, there is no (open source) implementation of an S3 source, so yes, you'll have to implement your own. You'll have to implement a Pollable source and the dev documentation has an outline that you can use. You can also look at the existing Execsource and work your way up. As far as

Re: how flume identifies a file transfer is complete or not

2014-07-26 Thread Sharninder
here any way to find that the transfer is > complete? > > Thanks, > Anand. > > > On 07/26/2014 01:38 PM, Sharninder wrote: > > If you really want to add files to HDFS, use the spool directory source > which is much more reliable. If you do want to use the exec source, n

Re: how flume identifies a file transfer is complete or not

2014-07-26 Thread Sharninder
If you really want to add files to HDFS, use the spool directory source which is much more reliable. If you do want to use the exec source, no point using cat since that's as good as cp'ing the file the HDFS, use tail -f rather. -- Sharninder On Sat, Jul 26, 2014 at 9:34 AM,

Re: Custom sink/source

2014-07-21 Thread Sharninder
I did look at the developer guide and was a little confused since there seem to be a couple of omissions in it. In particular, the even based source isn't explained very well. I'm new to Java (and flume) so maybe that's just me. Your explanation helps. -- Sharninder > On

Custom sink/source

2014-07-21 Thread Sharninder
will I be opening another connection again for the second event? When is stop() called? Has anyone implemented a websocket sink that I could take a look at? -- Sharninder

Re: Flume NG and S3

2014-06-30 Thread Sharninder
use-case. The reason the author might have moved to kafka, and I'm just speculating here, is that kafka provides him better buffering support for exactly the case I've written above. HTH Sharninder On Mon, Jun 30, 2014 at 7:57 PM, Máté Gulyás wrote: > Hi! > > I would l

End-to-End Encryption using flume

2014-06-25 Thread Sharninder
ble ? Hope my description was clear enough? -- Sharninder

Re: avro source file

2014-06-25 Thread Sharninder
oming from is not mentioned, or needed. -- Sharninder On Wed, Jun 25, 2014 at 5:03 PM, wrote: > My question is where is that input avro format file mentioned in the > configuration file. > > Sent from my iPhone > > On Jun 25, 2014, at 2:06 PM, Sharninder wrote: > >

Re: spooldir to hdfs

2014-06-25 Thread Sharninder
Did you try using the spooldir source with an hdfs sink? What problems did you face? -- Sharninder On Wed, Jun 25, 2014 at 12:15 PM, kishore alajangi < alajangikish...@gmail.com> wrote: > Hi Flume Experts, > > Could anybody help me to store avro files located in my localfilesys

Re: avro source file

2014-06-25 Thread Sharninder
That is an avro source, which means the producer or the app producing events needs to send it's data in the avro format. This is also used for building a multi-hop flume pipeline. Is this what you're asking? On Wed, Jun 25, 2014 at 12:37 PM, kishore alajangi < alajangikish...@gmail.com> wrote:

Re: copy to hdfs

2014-06-15 Thread Sharninder
;> >>> tier1.channels.channel1.capacity = 10* >>> tier1.sources.source1.deserializer.maxLineLength = 32768 >>> >>> >>> >>> the command I used is >>> >>> >>> >>> ./flume-ng agent --conf ./conf/ -f bin/e

Re: copy to hdfs

2014-06-15 Thread Sharninder
e, take a look at the spool directory source or the exec source and you should be able to put something together that'll push data through flume to hadoop. -- Sharninder

Re: Flume restart after sync destination failure

2014-06-13 Thread Sharninder
slightly unrelated but I've seen the flume file source also "hang" a few times on me. Restarting solved my problem. Can't really say more since I don't have a lot of data but flume failing on you is a problem and is the reason a lot of people develop their own "custom collectors". On a related not