[ 
https://issues.apache.org/jira/browse/IGNITE-2016?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15071500#comment-15071500
 ] 

Denis Magda commented on IGNITE-2016:
-------------------------------------

Roman,

According to {{SinkTask.put(...)}} API
http://docs.confluent.io/2.0.0/connect/javadocs/org/apache/kafka/connect/sink/SinkTask.html#put(java.util.Collection)
There is no need to buffer the data waiting for {{SinkTask.flush()}} to be 
called. It's ok and in my opinion preferable to start sending the data 
asynchronously.

I got your idea and concerns regarding the case when something going wrong 
during the streaming and we want to let Kafka know to continue streaming from 
some offset. However to overcome this issue we still can rely on 
{{IgniteDataStreamer}}. Just call {{IgniteDataStreamer.flush()}} in the 
implementation of your sink's flush method. This will ensure that the streamer 
has already flushed all the data properly at the time when implementation of 
your {{flush}} method returns.

Makes sense?

> Update KafkaStreamer to fit new features introduced in Kafka 0.9
> ----------------------------------------------------------------
>
>                 Key: IGNITE-2016
>                 URL: https://issues.apache.org/jira/browse/IGNITE-2016
>             Project: Ignite
>          Issue Type: New Feature
>          Components: streaming
>            Reporter: Roman Shtykh
>            Assignee: Roman Shtykh
>
> Particularly,
> - new consumer
> - Kafka Connect (Copycat)
> http://www.confluent.io/blog/apache-kafka-0.9-is-released
> This can be a a different integration task or a complete re-write of the 
> current implementation, considering the fact that Kafka Connect is a new 
> standard way for "large-scale, real-time data import and export for Kafka."



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to