Thanks Sudesh, for taking a look at this. I am already working on extending this for nanosecond precision as most sinks like iceberg have started giving nanosecond precision options.
On Wed, Apr 16, 2025 at 4:41 PM Sudesh Wasnik <wasnik...@gmail.com> wrote: > Hi Pritam ! Thanks for the KIP ! > Let’s extend the KIP to also add support for Nanosecond precision! > > Thanks > Sudesh > > On 2025/04/05 01:30:49 pritam kumar wrote: > > Hi Kafka Community, > > Sorry due to some changes I had to change the link to the kip. > > Here is the updated KIP link: > > > https://cwiki.apache.org/confluence/display/KAFKA/KIP-1154%3A+Extending+support+for+Microsecond+Precision+for+Kafka+Connect > > > > On Sat, Apr 5, 2025 at 12:14 AM pritam kumar <ku...@gmail.com> > > wrote: > > > > > Hi Kafka Community, > > > > > > I’d like to start a discussion on KIP-1153: Extending Support for > > > Microsecond Precision for Kafka Connect > > > < > https://cwiki.apache.org/confluence/display/KAFKA/KIP-1153%3A+Extending+Support+for+Microsecond+Precision+for+Kafka+Connect > > > > > . > > > > > > The primary motivation behind this KIP is to enhance the precision of > > > timestamp handling in Kafka Connect. Currently, Kafka Connect is > limited to > > > millisecond-level precision for timestamps. However, many modern data > > > formats and platforms have moved beyond this limitation: > > > > > > - > > > > > > Formats such as *Avro*, *Parquet*, and *ORC* support microsecond (and > > > even nanosecond) precision. For example, Avro specifies support for > > > timestamp-micros (spec link > > > <https://avro.apache.org/docs/1.11.0/spec.html#timestamp-micros>). > > > - > > > > > > Sink systems like *Apache Iceberg*, *Delta Lake*, and *Apache Hudi* > > > offer *microsecond and nanosecond precision* for time-based fields, > > > making millisecond precision inadequate for accurate data replication > and > > > analytics in some use cases. > > > > > > This gap can lead to *loss of precision* when transferring data through > > > Kafka Connect, especially when interacting with systems that depend on > > > high-resolution timestamps for change tracking, event ordering, or > > > deduplication. > > > > > > The goal of this KIP is to: > > > > > > - > > > > > > Introduce microsecond-level timestamp handling in Kafka Connect schema > > > and data representation. > > > - > > > > > > Ensure connectors (both source and sink) can leverage this precision > > > when supported by the underlying data systems. > > > - > > > > > > Maintain backward compatibility with existing millisecond-based > > > configurations and data. > > > > > > We welcome community feedback on: > > > > > > - > > > > > > Potential implementation concerns or edge cases we should address > > > - > > > > > > Suggestions for schema enhancements or conversion strategies > > > - > > > > > > Impacts on connector compatibility and testing > > > > > > Looking forward to your thoughts and input on this proposal! > > > > > > Thanks. > > > Link to the KIP. > > > > https://cwiki.apache.org/confluence/display/KAFKA/KIP-1153%3A+Extending+Support+for+Microsecond+Precision+for+Kafka+Connect > > > < > https://cwiki.apache.org/confluence/display/KAFKA/KIP-1152%3A+Add+transactional+ID+prefix+filter+to+ListTransactions+API > > > > > > > > > > Sent with a Spark >