Hi Pritam ! Thanks for the KIP !
Let’s extend the KIP to also add support for Nanosecond precision!

Thanks
Sudesh

On 2025/04/05 01:30:49 pritam kumar wrote:
> Hi Kafka Community,
> Sorry due to some changes I had to change the link to the kip.
> Here is the updated KIP link:
> https://cwiki.apache.org/confluence/display/KAFKA/KIP-1154%3A+Extending+support+for+Microsecond+Precision+for+Kafka+Connect
>
> On Sat, Apr 5, 2025 at 12:14 AM pritam kumar <ku...@gmail.com>
> wrote:
>
> > Hi Kafka Community,
> >
> > I’d like to start a discussion on KIP-1153: Extending Support for
> > Microsecond Precision for Kafka Connect
> > <https://cwiki.apache.org/confluence/display/KAFKA/KIP-1153%3A+Extending+Support+for+Microsecond+Precision+for+Kafka+Connect>
> > .
> >
> > The primary motivation behind this KIP is to enhance the precision of
> > timestamp handling in Kafka Connect. Currently, Kafka Connect is limited to
> > millisecond-level precision for timestamps. However, many modern data
> > formats and platforms have moved beyond this limitation:
> >
> > -
> >
> > Formats such as *Avro*, *Parquet*, and *ORC* support microsecond (and
> > even nanosecond) precision. For example, Avro specifies support for
> > timestamp-micros (spec link
> > <https://avro.apache.org/docs/1.11.0/spec.html#timestamp-micros>).
> > -
> >
> > Sink systems like *Apache Iceberg*, *Delta Lake*, and *Apache Hudi*
> > offer *microsecond and nanosecond precision* for time-based fields,
> > making millisecond precision inadequate for accurate data replication and
> > analytics in some use cases.
> >
> > This gap can lead to *loss of precision* when transferring data through
> > Kafka Connect, especially when interacting with systems that depend on
> > high-resolution timestamps for change tracking, event ordering, or
> > deduplication.
> >
> > The goal of this KIP is to:
> >
> > -
> >
> > Introduce microsecond-level timestamp handling in Kafka Connect schema
> > and data representation.
> > -
> >
> > Ensure connectors (both source and sink) can leverage this precision
> > when supported by the underlying data systems.
> > -
> >
> > Maintain backward compatibility with existing millisecond-based
> > configurations and data.
> >
> > We welcome community feedback on:
> >
> > -
> >
> > Potential implementation concerns or edge cases we should address
> > -
> >
> > Suggestions for schema enhancements or conversion strategies
> > -
> >
> > Impacts on connector compatibility and testing
> >
> > Looking forward to your thoughts and input on this proposal!
> >
> > Thanks.
> > Link to the KIP.
> > https://cwiki.apache.org/confluence/display/KAFKA/KIP-1153%3A+Extending+Support+for+Microsecond+Precision+for+Kafka+Connect
> > <https://cwiki.apache.org/confluence/display/KAFKA/KIP-1152%3A+Add+transactional+ID+prefix+filter+to+ListTransactions+API>
> >
> >

Sent with a Spark

Reply via email to