flink-connector-cassandra for C* version > 3.0

2018-06-04 Thread Meghashyam Sandeep V
Hi Guys, Has anyone used flink-connector-cassadra with cassandra version >= 3.X? I see that the cassandra version support is only 2.2.5 in the latest branch. https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-cassandra/pom.xml#L40 is there a way to pass these properties t

dynamic tables in cassandra sink

2018-05-03 Thread Meghashyam Sandeep V
Hi there, I have a flink stream from kafka writing to Cassandra. We use monthly tables in Cassandra to avoid TTL and tombstones that come with it. Tables would be like table_05_2018, table_06_2018 and so on. How do I dynamically register this table name in the following snippet? CassandraSink .ad

Re: Flink Kafka Consumer Behaviour

2017-04-24 Thread Meghashyam Sandeep V
Hi All, Sorry for the miscommunication. I'm not using 0.8. I'm using latest available flink-kafka client. I don't see my app registered as a consumer group. I wanted to know if there is a way to monitor Kafka offsets. Thanks, Sandeep On Apr 23, 2017 9:38 AM, "Stephan Ewen" wrote: > Since it is

Re: Suggestion for top 'k' products

2017-03-14 Thread Meghashyam Sandeep V
On Mon, Mar 13, 2017 at 11:22 AM, Meghashyam Sandeep V < vr1meghash...@gmail.com> wrote: > Thanks Suneel. Exactly what I was looking for. > > On Mon, Mar 13, 2017 at 10:31 AM, Suneel Marthi > wrote: > >> For an example implementation using Flink, check out >>

Re: Suggestion for top 'k' products

2017-03-13 Thread Meghashyam Sandeep V
nt in time. >> >> On Mon, Mar 13, 2017 at 1:25 PM, Meghashyam Sandeep V < >> vr1meghash...@gmail.com> wrote: >> >>> Hi All, >>> >>> I'm trying to use Flink for a use case where I would want to see my top >>> selling products in t

Suggestion for top 'k' products

2017-03-13 Thread Meghashyam Sandeep V
Hi All, I'm trying to use Flink for a use case where I would want to see my top selling products in time windows in near real time (windows of size 1-2 mins if fine). I guess this is the most common use case to use streaming apis in e-commerce. I see that I can iterate over records in a windowed s

Re: Amazon EMR Releases

2017-03-06 Thread Meghashyam Sandeep V
I spoke to one of the representatives in AWS EMR team last week. They mentioned that they usually practice a 4 week cool down period. Hopefully we will get Flink 1.2 in the next week. Thanks, Sandeep On Mar 6, 2017 9:27 AM, "Chen Qin" wrote: > EMR is a team within Amazon Web Services, to them,

Flink using notebooks in EMR

2017-03-02 Thread Meghashyam Sandeep V
Hi there, Has anyone tried using flink interpreter in Zeppelin using AWS EMR? I tried creating a new interpreter using host as 'localhots' and port '6123' which didn't seem to work. Thanks, Sandeep

Support for Auto scaling

2017-02-01 Thread Meghashyam Sandeep V
Hi Guys, I currently run flink 1.1.4 streaming jobs in EMR in AWS with yarn. I understand that EMR supports auto scaling but Flink doesn't. Is there a plan for this support in 1.2. Thanks, Sandeep

Re: benchmarking flink streaming

2017-01-25 Thread Meghashyam Sandeep V
amps, in order to measure the latency > of specific records. > > Stephan > > > On Fri, Dec 16, 2016 at 5:02 PM, Meghashyam Sandeep V < > vr1meghash...@gmail.com> wrote: > >> Hi Stephan, >> >> Thanks for your answer. Is there a way to get the metrics such a

Re: static/dynamic lookups in flink streaming

2016-12-21 Thread Meghashyam Sandeep V
t; operator, i.e., each operator holds a full copy of the map. > On the other hand, you do not need to shuffle the data because each > parallel task can do the look-up. > If your map is small, this would be the preferred approach. > > Best, Fabian > > 2016-12-21 18:46 GMT+01:00 Meghash

Re: static/dynamic lookups in flink streaming

2016-12-21 Thread Meghashyam Sandeep V
ink takes care of > checkpointing the state and restoring it in case of a failure. > In fact, the operator state is persisted in the state backends you > mentioned before. > > Best, Fabian > > 2016-12-21 15:02 GMT+01:00 Meghashyam Sandeep V : > >> Hi Fabian, >> >>

Re: static/dynamic lookups in flink streaming

2016-12-21 Thread Meghashyam Sandeep V
> I'm sorry but I think I do not understand your question. > What do you mean by static or dynamic look ups? Do you want to access an > external data store and cache data? > > Can you give a bit more detail about your use? > > Best, Fabian > > 2016-12-20 23:0

static/dynamic lookups in flink streaming

2016-12-20 Thread Meghashyam Sandeep V
Hi there, I know that there are various state backends to persist state. Is there a similar way to persist static/dynamic look ups and use them while streaming the data in Flink? Thanks, Sandeep

Re: How do I use values from a data stream to create a new streaming data source

2016-12-16 Thread Meghashyam Sandeep V
Runtime error is because you have non-serializable code in your 'map' operator. DataStream stockPrices = streamExecEnv.addSource(new LookupStockPrice(stockSymbol)); stockPrices.print(); The approach that you took will create infinite stockprice sources inside the ma

Re: benchmarking flink streaming

2016-12-16 Thread Meghashyam Sandeep V
ally be helpful to collect some jobs > in a form of "evaluation suite". > > Stephan > > > > On Thu, Dec 15, 2016 at 6:11 PM, Meghashyam Sandeep V < > vr1meghash...@gmail.com> wrote: > >> Hi There, >> >> We are evaluating Flink strea

benchmarking flink streaming

2016-12-15 Thread Meghashyam Sandeep V
Hi There, We are evaluating Flink streaming for real time data analysis. I have my flink job running in EMR with Yarn. What are the possible benchmarking tools that work best with Flink? I couldn't find this information in the Apache website. Thanks, Sandeep

Re: Multiple consumers and custom triggers

2016-12-15 Thread Meghashyam Sandeep V
1. If we have multiple sources, can the streams be parallelized ? 2. Can we have multiple sinks as well? On Dec 14, 2016 10:46 PM, wrote: > Got it. Thanks! > > On Dec 15, 2016, at 02:58, Jamie Grier wrote: > > Ahh, sorry, for #2: A single Flink job can have as many sources as you > like. They c

use case for sliding windows

2016-12-12 Thread Meghashyam Sandeep V
Hi There, I have a streaming job which has source as Kafka and sink as Cassandra. I have a use case where I wouldn't want to write some events to Cassandra when there are more than 100 events for a given 'id' (field in my Pojo) in 5mins. Is this a good usecase for SlidingWindows? Can I get the sli

Re: Reg. custom sinks in Flink

2016-12-12 Thread Meghashyam Sandeep V
eature to Flink's Cassandra Sink. > > Cheers, > Till > > On Mon, Dec 12, 2016 at 4:30 PM, Meghashyam Sandeep V < > vr1meghash...@gmail.com> wrote: > >> Data piles up in Cassandra without TTL. Is there a workaround for this >> problem? Is there a way to specify

Re: Reg. custom sinks in Flink

2016-12-12 Thread Meghashyam Sandeep V
pingManager. > We could add something similar as the ClusterBuilder for that though. > > Regards, > Chesnay > > > On 12.12.2016 16:15, Meghashyam Sandeep V wrote: > > Hi Till, > > Thanks for the information. > > 1. What do you mean by 'subtask', is i

Re: Reg. custom sinks in Flink

2016-12-12 Thread Meghashyam Sandeep V
ndra sink again. > > Cheers, > Till > ​ > > On Fri, Dec 9, 2016 at 8:05 PM, Meghashyam Sandeep V < > vr1meghash...@gmail.com> wrote: > >> Thanks a lot for the quick reply Shannon. >> >> 1. I will create a class that extends SinkFunction and write my >&

Re: Reg. custom sinks in Flink

2016-12-09 Thread Meghashyam Sandeep V
n > the position in or content in the stream. It's a necessary helper object, > yes, but you don't need Flink to checkpoint it. > > You can still use the sinks provided with flink-connector-cassandra and > customize the cluster building by passing your own ClusterBuilder into th

Reg. custom sinks in Flink

2016-12-09 Thread Meghashyam Sandeep V
Hi there, I have a flink streaming app where my source is Kafka and a custom sink to Cassandra(I can't use standard C* sink that comes with flink as I have customized auth to C*). I'm currently have the following: messageStream .rebalance() .map( s-> { return mapper.readValu