Hi,
I wrote a simple Flink job that uses Avro input format to read avro file
and save the results in avro format. The job does not get submitted and job
client exist out immediately. Same thing happens if I run the program in
IDE or if I submit via command line.
Here is the program-
import com.s
> On Wed, 23.03.2016 06:59, Chesnay Schepler wrote
> Could you be missing the call to execute()?
Yes, that was it. Can't believe I missed that !
Thank you Chesnay.
Best,
Tarandeep
On 23.03.2016 01:25, Tarandeep Singh wrote:
>> Hi,
>>
>> I wrote a simple Flink job t
Hi,
I am looking at implementation of zipWithIndex in DataSetUtils-
https://github.com/apache/flink/blob/master/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java
It works in two phases/steps
1) Count number of elements in each partition (using mapPartition)
2) In second m
Hi,
Can someone please point me to an example of creating DataSet using Avro
Generic Records?
I tried this code -
final ExecutionEnvironment env =
ExecutionEnvironment.getExecutionEnvironment();
final Path iPath = new Path(args[0]);
DataSet dataSet = env.createInput(new
AvroInputFor
che.org/jira/browse/FLINK-3691
>
> -Gna
>
> On Fri, Apr 1, 2016 at 4:04 AM, Tarandeep Singh
> wrote:
>
>> Hi,
>>
>> Can someone please point me to an example of creating DataSet using Avro
>> Generic Records?
>>
>> I tried this code -
>>
Hi,
How can I set compression for AvroOutputFormat when writing files on HDFS?
Also, can we set compression for intermediate data that is sent over
network (from map to reduce phase) ?
Thanks,
Tarandeep
ve some hints about which parts of the system would need to be
>> modified.
>>
>> – Ufuk
>>
>>
>> On Mon, Apr 18, 2016 at 12:56 PM, Tarandeep Singh
>> wrote:
>> > Hi,
>> >
>> > How can I set compression for AvroOu
Hi,
I have written ETL jobs in Flink (DataSet API). When I execute them in IDE,
they run and finish fine. When I try to run them on my cluster, I get
"Insufficient number of network buffers" error.
I have 5 machines in my cluster with 4 cores each. TaskManager is given 3GB
each. I increased the n
aded.
> I would suggest to increase the Akka ask timeouts via
> akka.ask.timeout: 100 s
> (
> https://ci.apache.org/projects/flink/flink-docs-release-1.0/setup/config.html#distributed-coordination-via-akka
> )
>
> – Ufuk
>
> On Tue, May 3, 2016 at 6:40 AM, Tarandeep Singh
Hi,
I am using DataSet API and reading Avro files as DataSet. I
am seeing this weird behavior that record is read correctly from file
(verified by printing all values) but when when this record is passed to
Flink chain/DAG (e.g. KeySelector), every field in this record has the same
value as the fi
On Wed, May 11, 2016 at 10:24 PM, Tarandeep Singh
wrote:
> Hi,
>
> I am using DataSet API and reading Avro files as DataSet. I
> am seeing this weird behavior that record is read correctly from file
> (verified by printing all values) but when when this record is passed to
> Fl
Hi,
I came across Flink stateful functions and the project/idea excited me. But I
noticed no code has been committed lately in this project and I couldn’t fins
it’s roadmap.
Can someone shed some light on state if this project?
Thank you,
Tarandeep
Hi,
I am getting NoSerializableException in this class-
public class RecordsFilterer {
public DataSet> addFilterFlag(DataSet
dataset, DataSet filteredIds, String fieldName) {
return dataset.coGroup(filteredIds)
.where(new KeySelector() {
@Overri
ot;myField");
>>
>> Regarding the join code, I would suggest an optimization.
>> Instead of using CoGroup, I would use distinct and an OuterJoin like this:
>>
>> DataSet distIds = filtereredIds.distinct();
>> DataSet result = records
>> .leftOuterJ
Hi,
I am using flink-1.0.0 and running ETL (batch) jobs on it for quite some
time (few months) without any problem. Starting this morning, I have been
getting errors like these-
"Received an event in channel 3 while still having data from a record. This
indicates broken serialization logic. If yo
com.esotericsoftware.kryo.util.DefaultClassResolver.readName(DefaultClassResolver.java:136)
... 15 more
-Tarandeep
On Mon, Oct 3, 2016 at 12:49 PM, Tarandeep Singh
wrote:
> Hi,
>
> I am using flink-1.0.0 and running ETL (batch) jobs on it for quite some
> time (few months) without any problem. Starting this morning, I have been
>
>> wrote:
> >>>
> >>> I think you're running into the same exception I face sometimes..I've
> >>> opened a jira for it [1]. Could you please try to apply that patch and
> see
> >>> if things get better?
> >>>
> >>&g
Hi,
Looking forward to 1.2 version of Flink (lots of exciting features have
been added).
Has the date finalized yet?
Thanks,
Tarandeep
Hi,
I am running a flink cluster to process clickstream data (generate user
level, page level, site level statistics)
I want to understand the cons and pros of submitting multiple jobs (each
job handles one simple processing/computation) vs one/few complex jobs. At
present, the events are read
Hi,
Our flink streaming job that is reading from old version of Kafka keeps
failing (every 9 minutes or so) with this error:
java.lang.RuntimeException: Unable to retrieve any partitions for the
requested topics [extracted-dimensions]. Please check previous log entries
at
org.apache.flink
Hi Robert & Nico,
I am facing the same problem (java.lang.NoClassDefFoundError:
com/codahale/metrics/Metric)
Can you help me identify shading issue in pom.xml file.
My pom.xml content-
-
http://maven.apache.org/POM/4.0.0";
xmlns:xsi="http://www.w3.org
Hi,
I am using flink-1.2 streaming API to process clickstream and compute some
results per cookie. The computed results are stored in Cassandra using
flink-cassandra connector. After a result is stored in cassandra, I want to
notify an external system (using their API or via Kafka) that result is
Hi,
I am using flink-1.2 and reading data stream from Kafka (using
FlinkKafkaConsumer08). I want to connect this data stream with another
stream (read control stream) so as to do some filtering on the fly. After
filtering, I am applying window function (tumbling/sliding event window)
along with fo
}
>
>
> Have you tried using a different “group.id” everytime you’re re-running the
> job?
> Note that the “auto.offset.reset” value is only respected when there aren’t
> any offsets for the group committed in Kafka.
> So you might not actually be reading the complete “small_inp
Anyone?
Any suggestions what could be going wrong or what I am doing wrong?
Thanks,
Tarandeep
On Thu, Mar 16, 2017 at 7:34 AM, Tarandeep Singh
wrote:
> Data is read from Kafka and yes I use different group id every time I run
> the code. I have put break points and print statements to
in your job. Have you
> checked whether or not the watermarks printed there are identical (using
> getInput v.s. getKafkaInput)?
>
> Cheers,
> Gordon
>
>
> On March 17, 2017 at 12:32:51 PM, Tarandeep Singh (tarand...@gmail.com)
> wrote:
>
> Anyone?
> Any suggesti
Hi,
Is someone using flinkspector unit testing framework with flink-1.2?
I added the following dependencies in my pom.xml file:
org.flinkspector
flinkspector-datastream_2.10
0.5
org.flinkspector
flinkspector-co
ote:
> Can you post the patch for flink-specter where the mini cluster is
> replaced ?
>
> I assume you upgraded the version of Flink in the pom.
>
> Cheers
>
> On Mar 17, 2017, at 4:26 PM, Tarandeep Singh wrote:
>
> Hi,
>
> Is someone using flinkspector unit test
>
> Cheers
>
> On Fri, Mar 17, 2017 at 5:19 PM, Tarandeep Singh
> wrote:
>
>> Hi Ted,
>>
>> See the attached patch.
>>
>> I am able to run test examples (e.g.
>> org.flinkspector.datastream.examples.TestMapper)
>> via IntelliJ. But when
Hi Nancy,
I also get 1 test failed when I build/run tests on flink-spector:
- should stop if all triggers fire
Run completed in 3 seconds, 944 milliseconds.
Total number of tests run: 19
Suites: completed 5, aborted 0
Tests: succeeded 18, failed 1, canceled 0, ignored 0, pending 0
*** 1 TEST FAIL
Hi,
I am using flink-1.2 and Cassandra connector to write to cassandra tables.
I am using POJOs with DataStax annotations as described here-
https://ci.apache.org/projects/flink/flink-docs-release-1.3/dev/connectors/cassandra.html
My question is- how are nulls handles by cassandra sink?
Datastax
you to cherry-pick it onto a 1.2 branch.
>
> I will add a ticket for this soon (currently getting timeouts in JIRA).
>
> Regards,
> Chesnay
>
>
> On 12.04.2017 02:27, Tarandeep Singh wrote:
>
>> Hi,
>>
>> I am using flink-1.2 and Cassandra connector to w
Hi,
Any updates on 1.3 release date?
Thanks,
Tarandeep
Hi Chesnay,
Did your code changes (exposing mapper options) made it in 1.3 release?
Thank you,
Tarandeep
On Wed, Apr 12, 2017 at 2:34 PM, Tarandeep Singh
wrote:
> Thanks Chesnay, this will work.
>
> Best,
> Tarandeep
>
> On Wed, Apr 12, 2017 at 2:42 AM, Chesnay Schepler
&
No problem :)
Thanks for letting me know.
Best,
Tarandeep
On Thu, Jun 1, 2017 at 11:18 AM, Chesnay Schepler
wrote:
> No, unfortunately I forgot about them :/
>
>
> On 01.06.2017 19:39, Tarandeep Singh wrote:
>
> Hi Chesnay,
>
> Did your code changes (exposing mapper
35 matches
Mail list logo