Yes, the Flink Kinesis Consumer detects aggregated records and seamlessly de-aggregates them for you [1].
Thanks, [1] https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/internals/publisher/RecordBatch.java#L93 On Thu, 14 Apr 2022, 23:56 Blake Wilson, <bl...@yellowpapersun.com> wrote: > Thanks for offering to review, Danny. > > Thanks also for pointing out that KCL can de-aggregate records aggregated > by KPL. Several applications I've worked on batch multiple records without > using the KPL unfortunately. > > Is de-aggregation supported by the Kinesis Connector Source? I found > mention of aggregation only in the FlinkKinesisProducer when searching > online for this feature. > > On Thu, Apr 14, 2022 at 12:51 AM Danny Cranmer <dannycran...@apache.org> > wrote: > > > Just to clarify, the native KCL/KPL aggregation [1] handles the partition > > key rebalancing for you out of the box. > > > > > > [1] https://docs.aws.amazon.com/streams/latest/dev/kinesis > > -kpl-concepts.html#kinesis-kpl-concepts-aggretation > > > > On Thu, Apr 14, 2022 at 8:48 AM Danny Cranmer <dannycran...@apache.org> > > wrote: > > > > > Hey Blake, > > > > > > I am happy to take a look, but I will not have capacity until next > week. > > > > > > The current way to achieve multiple records per PUT is to use the > native > > > KCL/KPL aggregation [1], which is supported by the Flink connector. A > > > downside of aggregation is that the sender has to manage the > partitioning > > > strategy. For example, each record in your list will be sent to the > same > > > shard. If the sender implements grouping of records by partition key, > > then > > > care needs to be taken during shard scaling. > > > > > > Thanks, > > > > > > [1] > > > > > > https://docs.aws.amazon.com/streams/latest/dev/kinesis-kpl-concepts.html#kinesis-kpl-concepts-aggretation > > > > > > > > > On Tue, Apr 12, 2022 at 3:52 AM Blake Wilson <bl...@yellowpapersun.com > > > > > wrote: > > > > > >> Hello, I recently submitted a pull request to support the Collector > API > > >> for > > >> the Kinesis Streams Connector. > > >> > > >> The ability to use this API would save a great deal of shuttling bytes > > >> around in multiple Flink programs I've worked on. This is because to > > >> construct a stream of the desired type without Collector support, the > > >> Kinesis source must emit a List[Type], and this must be flattened to a > > >> Type > > >> stream. > > >> > > >> Because of the way Kinesis pricing works, it rarely makes sense to > send > > >> one > > >> value per Kinesis record. In provisioned mode, Kinesis PUTs are priced > > to > > >> the nearest 25KB ( > https://aws.amazon.com/kinesis/data-streams/pricing/ > > ), > > >> so > > >> records are more sensibly packed with multiple values unless these > > values > > >> are quite large. Therefore, I suspect the need to handle multiple > values > > >> per Kinesis record is quite common. > > >> > > >> The PR is located at https://github.com/apache/flink/pull/19417, and > > I'd > > >> love to get some feedback on Github or here. > > >> > > >> Thanks! > > >> > > > > > >