Nick,

There will be drawbacks to this because you would be using the wrong tool
for the task. The Kafka ecosystem is very rich with different tools and you
should spend a little more time to understand what they do and are meant for

I believe the ecosystem component you are looking for is Kafka Streams. You
can also use KSQLDB for this if it is simple enough and you want to avoid
writing Java/Scala code

You should spend some time to understand each of the components of the
ecosystem and it will become clearer how they can be leveraged for stream
processing:

https://kafka.apache.org/documentation/#api

Kafka Brokers
Kafka Connect
Kafka Streams
Kafka Producers
Kafka Consumers
Kafka Admin Clients

Confluent Schema Registry
Confluent REST Proxy
Confluent KSQLDB

Take a look at all of the above and try to understand how they are inter
connected, dependent and used and it will bring better clarity

To directly answer your question, you do not need a connector to process
your events, you need to create a Kafka Streams app or depending on the use
case leverage KSQLDB for the task. The number of instances can scale up or
down as needed. The connect framework was not intended for what you are
attempting to use it for.

Please follow up with any additional questions you may have

I hope this was helpful


On Thu, Mar 11, 2021 at 1:27 PM Nick Siviglia <nisivig...@gmail.com> wrote:

> Hi Everyone,
>
> I'd like to create a Kafka connector that instead of acting like a source
> or sink will instead do some processing on the data and output to another
> kafka topic. Has anyone done this before? Does anyone see any potential
> drawbacks?
>
> Data is json format at around 150 string and number fields per object. And
> I'm planning on receiving about 2 million a day.
>
> Thanks for any help,
> Nick
>

Reply via email to