Unable to submit flink job that uses Avro data

2016-03-22 Thread Tarandeep Singh
Hi, I wrote a simple Flink job that uses Avro input format to read avro file and save the results in avro format. The job does not get submitted and job client exist out immediately. Same thing happens if I run the program in IDE or if I submit via command line. Here is the program- import com.s

Re: Unable to submit flink job that uses Avro data

2016-03-23 Thread Tarandeep Singh
> On Wed, 23.03.2016 06:59, Chesnay Schepler wrote > Could you be missing the call to execute()? Yes, that was it. Can't believe I missed that ! Thank you Chesnay. Best, Tarandeep On 23.03.2016 01:25, Tarandeep Singh wrote: >> Hi, >> >> I wrote a simple Flink job t

DataSetUtils zipWithIndex question

2016-03-29 Thread Tarandeep Singh
Hi, I am looking at implementation of zipWithIndex in DataSetUtils- https://github.com/apache/flink/blob/master/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java It works in two phases/steps 1) Count number of elements in each partition (using mapPartition) 2) In second m

Example - Reading Avro Generic records

2016-04-01 Thread Tarandeep Singh
Hi, Can someone please point me to an example of creating DataSet using Avro Generic Records? I tried this code - final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment(); final Path iPath = new Path(args[0]); DataSet dataSet = env.createInput(new AvroInputFor

Re: Example - Reading Avro Generic records

2016-04-02 Thread Tarandeep Singh
che.org/jira/browse/FLINK-3691 > > -Gna > > On Fri, Apr 1, 2016 at 4:04 AM, Tarandeep Singh > wrote: > >> Hi, >> >> Can someone please point me to an example of creating DataSet using Avro >> Generic Records? >> >> I tried this code - >>

Compression - AvroOutputFormat and over network ?

2016-04-18 Thread Tarandeep Singh
Hi, How can I set compression for AvroOutputFormat when writing files on HDFS? Also, can we set compression for intermediate data that is sent over network (from map to reduce phase) ? Thanks, Tarandeep

Re: Compression - AvroOutputFormat and over network ?

2016-04-20 Thread Tarandeep Singh
ve some hints about which parts of the system would need to be >> modified. >> >> – Ufuk >> >> >> On Mon, Apr 18, 2016 at 12:56 PM, Tarandeep Singh >> wrote: >> > Hi, >> > >> > How can I set compression for AvroOu

Insufficient number of network buffers

2016-05-02 Thread Tarandeep Singh
Hi, I have written ETL jobs in Flink (DataSet API). When I execute them in IDE, they run and finish fine. When I try to run them on my cluster, I get "Insufficient number of network buffers" error. I have 5 machines in my cluster with 4 cores each. TaskManager is given 3GB each. I increased the n

Re: Insufficient number of network buffers

2016-05-03 Thread Tarandeep Singh
aded. > I would suggest to increase the Akka ask timeouts via > akka.ask.timeout: 100 s > ( > https://ci.apache.org/projects/flink/flink-docs-release-1.0/setup/config.html#distributed-coordination-via-akka > ) > > – Ufuk > > On Tue, May 3, 2016 at 6:40 AM, Tarandeep Singh

Flink + Avro GenericRecord - first field value overwrites all other fields

2016-05-11 Thread Tarandeep Singh
Hi, I am using DataSet API and reading Avro files as DataSet. I am seeing this weird behavior that record is read correctly from file (verified by printing all values) but when when this record is passed to Flink chain/DAG (e.g. KeySelector), every field in this record has the same value as the fi

Re: Flink + Avro GenericRecord - first field value overwrites all other fields

2016-05-12 Thread Tarandeep Singh
On Wed, May 11, 2016 at 10:24 PM, Tarandeep Singh wrote: > Hi, > > I am using DataSet API and reading Avro files as DataSet. I > am seeing this weird behavior that record is read correctly from file > (verified by printing all values) but when when this record is passed to > Fl

Stateful functions roadmap

2023-06-01 Thread Tarandeep Singh
Hi, I came across Flink stateful functions and the project/idea excited me. But I noticed no code has been committed lately in this project and I couldn’t fins it’s roadmap. Can someone shed some light on state if this project? Thank you, Tarandeep

NotSerializableException

2016-06-08 Thread Tarandeep Singh
Hi, I am getting NoSerializableException in this class-  public class RecordsFilterer { public DataSet> addFilterFlag(DataSet dataset, DataSet filteredIds, String fieldName) { return dataset.coGroup(filteredIds) .where(new KeySelector() { @Overri

Re: NotSerializableException

2016-06-09 Thread Tarandeep Singh
ot;myField"); >> >> Regarding the join code, I would suggest an optimization. >> Instead of using CoGroup, I would use distinct and an OuterJoin like this: >> >> DataSet distIds = filtereredIds.distinct(); >> DataSet result = records >> .leftOuterJ

Exception while running Flink jobs (1.0.0)

2016-10-03 Thread Tarandeep Singh
Hi, I am using flink-1.0.0 and running ETL (batch) jobs on it for quite some time (few months) without any problem. Starting this morning, I have been getting errors like these- "Received an event in channel 3 while still having data from a record. This indicates broken serialization logic. If yo

Re: Exception while running Flink jobs (1.0.0)

2016-10-03 Thread Tarandeep Singh
com.esotericsoftware.kryo.util.DefaultClassResolver.readName(DefaultClassResolver.java:136) ... 15 more -Tarandeep On Mon, Oct 3, 2016 at 12:49 PM, Tarandeep Singh wrote: > Hi, > > I am using flink-1.0.0 and running ETL (batch) jobs on it for quite some > time (few months) without any problem. Starting this morning, I have been >

Re: Exception while running Flink jobs (1.0.0)

2016-10-05 Thread Tarandeep Singh
>> wrote: > >>> > >>> I think you're running into the same exception I face sometimes..I've > >>> opened a jira for it [1]. Could you please try to apply that patch and > see > >>> if things get better? > >>> > >>&g

1.2 release date

2017-02-05 Thread Tarandeep Singh
Hi, Looking forward to 1.2 version of Flink (lots of exciting features have been added). Has the date finalized yet? Thanks, Tarandeep

Many streaming jobs vs one

2017-02-05 Thread Tarandeep Singh
Hi, I am running a flink cluster to process clickstream data (generate user level, page level, site level statistics) I want to understand the cons and pros of submitting multiple jobs (each job handles one simple processing/computation) vs one/few complex jobs. At present, the events are read

Flink streaming (1.3.2) KafkaConsumer08 - Unable to retrieve any partitions

2018-01-22 Thread Tarandeep Singh
Hi, Our flink streaming job that is reading from old version of Kafka keeps failing (every 9 minutes or so) with this error: java.lang.RuntimeException: Unable to retrieve any partitions for the requested topics [extracted-dimensions]. Please check previous log entries at org.apache.flink

Re: Flink 1.2 and Cassandra Connector

2017-03-06 Thread Tarandeep Singh
Hi Robert & Nico, I am facing the same problem (java.lang.NoClassDefFoundError: com/codahale/metrics/Metric) Can you help me identify shading issue in pom.xml file. My pom.xml content- - http://maven.apache.org/POM/4.0.0"; xmlns:xsi="http://www.w3.org

Flink streaming - call external API "after" sink

2017-03-09 Thread Tarandeep Singh
Hi, I am using flink-1.2 streaming API to process clickstream and compute some results per cookie. The computed results are stored in Cassandra using flink-cassandra connector. After a result is stored in cassandra, I want to notify an external system (using their API or via Kafka) that result is

Data+control stream from kafka + window function - not working

2017-03-15 Thread Tarandeep Singh
Hi, I am using flink-1.2 and reading data stream from Kafka (using FlinkKafkaConsumer08). I want to connect this data stream with another stream (read control stream) so as to do some filtering on the fly. After filtering, I am applying window function (tumbling/sliding event window) along with fo

Re: Data+control stream from kafka + window function - not working

2017-03-16 Thread Tarandeep Singh
} > > > Have you tried using a different “group.id” everytime you’re re-running the > job? > Note that the “auto.offset.reset” value is only respected when there aren’t > any offsets for the group committed in Kafka. > So you might not actually be reading the complete “small_inp

Re: Data+control stream from kafka + window function - not working

2017-03-16 Thread Tarandeep Singh
Anyone? Any suggestions what could be going wrong or what I am doing wrong? Thanks, Tarandeep On Thu, Mar 16, 2017 at 7:34 AM, Tarandeep Singh wrote: > Data is read from Kafka and yes I use different group id every time I run > the code. I have put break points and print statements to

Re: Data+control stream from kafka + window function - not working

2017-03-17 Thread Tarandeep Singh
in your job. Have you > checked whether or not the watermarks printed there are identical (using > getInput v.s. getKafkaInput)? > > Cheers, > Gordon > > > On March 17, 2017 at 12:32:51 PM, Tarandeep Singh (tarand...@gmail.com) > wrote: > > Anyone? > Any suggesti

flink-1.2 and unit testing / flinkspector

2017-03-17 Thread Tarandeep Singh
Hi, Is someone using flinkspector unit testing framework with flink-1.2? I added the following dependencies in my pom.xml file: org.flinkspector flinkspector-datastream_2.10 0.5 org.flinkspector flinkspector-co

Re: flink-1.2 and unit testing / flinkspector

2017-03-17 Thread Tarandeep Singh
ote: > Can you post the patch for flink-specter where the mini cluster is > replaced ? > > I assume you upgraded the version of Flink in the pom. > > Cheers > > On Mar 17, 2017, at 4:26 PM, Tarandeep Singh wrote: > > Hi, > > Is someone using flinkspector unit test

Re: flink-1.2 and unit testing / flinkspector

2017-03-17 Thread Tarandeep Singh
> > Cheers > > On Fri, Mar 17, 2017 at 5:19 PM, Tarandeep Singh > wrote: > >> Hi Ted, >> >> See the attached patch. >> >> I am able to run test examples (e.g. >> org.flinkspector.datastream.examples.TestMapper) >> via IntelliJ. But when

Re: flink-1.2 and unit testing / flinkspector

2017-03-23 Thread Tarandeep Singh
Hi Nancy, I also get 1 test failed when I build/run tests on flink-spector: - should stop if all triggers fire Run completed in 3 seconds, 944 milliseconds. Total number of tests run: 19 Suites: completed 5, aborted 0 Tests: succeeded 18, failed 1, canceled 0, ignored 0, pending 0 *** 1 TEST FAIL

Cassandra connector POJO - tombstone question

2017-04-11 Thread Tarandeep Singh
Hi, I am using flink-1.2 and Cassandra connector to write to cassandra tables. I am using POJOs with DataStax annotations as described here- https://ci.apache.org/projects/flink/flink-docs-release-1.3/dev/connectors/cassandra.html My question is- how are nulls handles by cassandra sink? Datastax

Re: Cassandra connector POJO - tombstone question

2017-04-12 Thread Tarandeep Singh
you to cherry-pick it onto a 1.2 branch. > > I will add a ticket for this soon (currently getting timeouts in JIRA). > > Regards, > Chesnay > > > On 12.04.2017 02:27, Tarandeep Singh wrote: > >> Hi, >> >> I am using flink-1.2 and Cassandra connector to w

Flink 1.3 release date

2017-06-01 Thread Tarandeep Singh
Hi, Any updates on 1.3 release date? Thanks, Tarandeep

Re: Cassandra connector POJO - tombstone question

2017-06-01 Thread Tarandeep Singh
Hi Chesnay, Did your code changes (exposing mapper options) made it in 1.3 release? Thank you, Tarandeep On Wed, Apr 12, 2017 at 2:34 PM, Tarandeep Singh wrote: > Thanks Chesnay, this will work. > > Best, > Tarandeep > > On Wed, Apr 12, 2017 at 2:42 AM, Chesnay Schepler &

Re: Cassandra connector POJO - tombstone question

2017-06-01 Thread Tarandeep Singh
No problem :) Thanks for letting me know. Best, Tarandeep On Thu, Jun 1, 2017 at 11:18 AM, Chesnay Schepler wrote: > No, unfortunately I forgot about them :/ > > > On 01.06.2017 19:39, Tarandeep Singh wrote: > > Hi Chesnay, > > Did your code changes (exposing mapper