[ 
https://issues.apache.org/jira/browse/FLINK-35808?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Lam updated FLINK-35808:
------------------------------
    Description: 
This issue is a follow-up to [this mailing list 
discussion|https://lists.apache.org/thread/spl88o63sjm2dv4l5no0ym632d2yt2o6]. 

I'd like to propose letting the ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG 
be overridable by user in KafkaSourceBuilder, as shown in this DRAFT PR:
 
[https://github.com/apache/flink-connector-kafka/pull/108]
 
>From the PR description: 
{quote}This allows users to easily implement the [{{claim check}} large message 
pattern|https://developer.confluent.io/patterns/event-processing/claim-check/] 
without bringing any concerns into the Flink codebase otherwise, by specifying 
a {{value.deserializer}} that handles it, but otherwise passes through the 
bytes.
Note: [overriding {{value.serializer}} is already supported on the Producer 
side.](
|[https://github.com/apache/flink-connector-kafka/blob/15d3fbd4e65dae6c334e2386dd337d2bf423c216/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaSinkBuilder.java#L82-L83|https://github.com/apache/flink-connector-kafka/blob/15d3fbd4e65dae6c334e2386dd337d2bf423c216/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaSinkBuilder.java#L82-L83]])|
|
 
Other Reading:
[https://www.kai-waehner.de/blog/2020/08/07/apache-kafka-handling-large-messages-and-files-for-image-video-audio-processing/
] 
[https://www.conduktor.io/kafka/how-to-send-large-messages-in-apache-kafka/#Option-1:-using-an-external-store-(GB-size-messages)-0]|
{quote}
 
What do folks think? If it seems reasonable I can follow the steps in the 
[contribution guide|https://flink.apache.org/how-to-contribute/overview/].  

  was:
This issue is a follow-up to [this mailing list 
discussion|https://lists.apache.org/thread/spl88o63sjm2dv4l5no0ym632d2yt2o6]. 

I'd like to propose letting the ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG 
be overridable by user in KafkaSourceBuilder, as shown in this DRAFT PR:
 
https://github.com/apache/flink-connector-kafka/pull/108
 
>From the PR description: 
{quote}This allows users to easily implement the [{{claim check}} large message 
pattern|https://developer.confluent.io/patterns/event-processing/claim-check/] 
without bringing any concerns into the Flink codebase otherwise, by specifying 
a {{value.deserializer}} that handles it, but otherwise passes through the 
bytes.
Note: [overriding {{value.serializer}} is already supported on the Producer 
side.
|https://github.com/apache/flink-connector-kafka/blob/15d3fbd4e65dae6c334e2386dd337d2bf423c216/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaSinkBuilder.java#L82-L83]Other
 Reading:
[https://www.kai-waehner.de/blog/2020/08/07/apache-kafka-handling-large-messages-and-files-for-image-video-audio-processing/
] 
[https://www.conduktor.io/kafka/how-to-send-large-messages-in-apache-kafka/#Option-1:-using-an-external-store-(GB-size-messages)-0]{quote}
 
What do folks think? If it seems reasonable I can follow the steps in the 
[contribution guide|https://flink.apache.org/how-to-contribute/overview/].  


> Let ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG be overridable by user in 
> KafkaSourceBuilder
> -----------------------------------------------------------------------------------------------
>
>                 Key: FLINK-35808
>                 URL: https://issues.apache.org/jira/browse/FLINK-35808
>             Project: Flink
>          Issue Type: Improvement
>          Components: Connectors / Kafka
>    Affects Versions: kafka-3.2.0
>            Reporter: Kevin Lam
>            Priority: Minor
>
> This issue is a follow-up to [this mailing list 
> discussion|https://lists.apache.org/thread/spl88o63sjm2dv4l5no0ym632d2yt2o6]. 
> I'd like to propose letting the 
> ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG be overridable by user in 
> KafkaSourceBuilder, as shown in this DRAFT PR:
>  
> [https://github.com/apache/flink-connector-kafka/pull/108]
>  
> From the PR description: 
> {quote}This allows users to easily implement the [{{claim check}} large 
> message 
> pattern|https://developer.confluent.io/patterns/event-processing/claim-check/]
>  without bringing any concerns into the Flink codebase otherwise, by 
> specifying a {{value.deserializer}} that handles it, but otherwise passes 
> through the bytes.
> Note: [overriding {{value.serializer}} is already supported on the Producer 
> side.](
> |[https://github.com/apache/flink-connector-kafka/blob/15d3fbd4e65dae6c334e2386dd337d2bf423c216/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaSinkBuilder.java#L82-L83|https://github.com/apache/flink-connector-kafka/blob/15d3fbd4e65dae6c334e2386dd337d2bf423c216/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaSinkBuilder.java#L82-L83]])|
> |
>  
> Other Reading:
> [https://www.kai-waehner.de/blog/2020/08/07/apache-kafka-handling-large-messages-and-files-for-image-video-audio-processing/
> ] 
> [https://www.conduktor.io/kafka/how-to-send-large-messages-in-apache-kafka/#Option-1:-using-an-external-store-(GB-size-messages)-0]|
> {quote}
>  
> What do folks think? If it seems reasonable I can follow the steps in the 
> [contribution guide|https://flink.apache.org/how-to-contribute/overview/].  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to