RE: Wordindex conversation.

2016-10-10 Thread Kürşat Kurt
Ok, thanks Fabian. From: Fabian Hueske [mailto:fhue...@gmail.com] Sent: Tuesday, October 11, 2016 1:12 AM To: user@flink.apache.org Subject: Re: Wordindex conversation. Hi, you can do it like this: 1) you have to split each label record of the main dataset into separate records: (0

Re: Wordindex conversation.

2016-10-10 Thread Fabian Hueske
Hi, you can do it like this: 1) you have to split each label record of the main dataset into separate records: (0,List(a, b, c, d, e, f, g)) -> (0, a), (0, b), (0, c), ..., (0, g) (1,List(b, c, f, a, g)) -> (1, b), (1, c), ..., (1, g) 2) join word index dataset with splitted main dataset: Data

Wordindex conversation.

2016-10-10 Thread Kürşat Kurt
Hi; I have MainDataset (Label,WordList) : (0,List(a, b, c, d, e, f, g)) (1,List(b, c, f, a, g)) ..and, wordIndex dataset(created with .zipWithIndex) : wordIndex> (0,a) wordIndex> (1,b) wordIndex> (2,c) wordIndex> (3,d) wordIndex> (4,e) wordIndex> (5,f) wordIndex> (6,g) H

Re: jdbc.JDBCInputFormat

2016-10-10 Thread Timo Walther
I could reproduce the error locally. I will prepare a fix for it. Timo Am 10/10/16 um 11:54 schrieb Alberto Ramón: It's from Jun and Unassigned :( Is There a Workarround? I'm will try to contact with the reporter , Martin Scholl ) 2016-10-10 11:04 GMT+02:00 Timo Walther

Re: Processing events through web socket

2016-10-10 Thread Maximilian Michels
You don't have to wrap the existing socket source. You will have to create your own source function which directly reads from the web socket and forwards the results to the next Flink operator. Then you will call env.addSource(new YourWebSocketSource()); To get started, you might want to read the

Re: Iterative queries on Flink

2016-10-10 Thread Flavio Pompermaier
Thaks Stephan for the answer. As I told to Fabian we need to apply some transformation to datasets interactively. For the moment I will use livy + spark[1] but I'll prefer to stick with Flink if possible. So, if there's any effor in this direction just let me know and I'll be happy to contribute.

Re: Iterative queries on Flink

2016-10-10 Thread Stephan Ewen
There is still quite a bit needed to do this properly: (1) incremental recovery (2) network stack caching (1) will probably happen quite soon, I am not aware of any committer having concrete plans for (2). Best, Stephan On Sat, Oct 8, 2016 at 4:41 PM, Flavio Pompermaier wrote: > Any progr

Re: Executing a flink program

2016-10-10 Thread Maximilian Michels
Whoops, I meant Flink and not Beam (I had just visited the Beam mailing list). -Max On Mon, Oct 10, 2016 at 12:08 PM, Maximilian Michels wrote: > Normally, you should be able to directly execute your Beam program > from within your IDE. It automatically starts a local cluster with the > resourc

Re: Executing a flink program

2016-10-10 Thread Maximilian Michels
Normally, you should be able to directly execute your Beam program from within your IDE. It automatically starts a local cluster with the resources needed for the job. Which Beam version are you using? Could you post some of the code your executing? -Max On Sat, Oct 8, 2016 at 7:51 PM, Dayong

Re: jdbc.JDBCInputFormat

2016-10-10 Thread Alberto Ramón
It's from Jun and Unassigned :( Is There a Workarround? I'm will try to contact with the reporter , Martin Scholl ) 2016-10-10 11:04 GMT+02:00 Timo Walther : > I think you already found the correct issue describing your problem ( > FLINK-4108). This should get higher priority. > > Timo > > Am

Re: jdbc.JDBCInputFormat

2016-10-10 Thread Timo Walther
I think you already found the correct issue describing your problem ( FLINK-4108). This should get higher priority. Timo Am 09/10/16 um 13:27 schrieb Alberto Ramón: After solved some issues, I connected with Kylin, but I can't read data import org.apache.flink.api.scala._ import org.apache.f

Re: Handling decompression exceptions

2016-10-10 Thread Yassine MARZOUGUI
Thank you Fabian and Stephan for the suggestions. I couldn't override "readLine()" because it's final, so went with Fabian's solution, but I'm struggling with csv field masks. Any help is appreciated. I created an Input Format which is basically TupleCsvInputFormat for which I overrode the nextReco