Hi Karthi
Thanks for the proposal.
I personally like the idea of having a specific prometheus sink vs using
metrics reporters. I think this is a clear boundary between monitoring the
job metrics and treating metrics as data payloads.
Additionally we would benefit of AsyncSink framework :D.

I would love to reiterate on Danny's point regarding credentials providers
and authentication, and on Jing's question regarding source if we have it
in plan.
I would also assume we will be adding support to Table API connector, can
we clarify that in the FLIP.

Best Regards,
Ahmed Hamdy


On Fri, 19 May 2023 at 07:41, Yun Tang <myas...@live.com> wrote:

> +1 for the idea to introduce the prometheus connector.
>
> In some cases, users cannot leverage the current way to report very large
> amounts of metrics via metrics reporter due to the limit on the message
> collection of prometheus agent.
> For the FLIP itself, I think we should also consider the rate limit
> feature to avoid breaking down the metrics services.
>
> Best
> Yun Tang
> ________________________________
> From: Jing Ge <j...@ververica.com.INVALID>
> Sent: Friday, May 19, 2023 4:49
> To: dev@flink.apache.org <dev@flink.apache.org>
> Subject: Re: [DISCUSS] FLIP-229: Prometheus Sink Connector
>
> Hi Karthi,
>
> Thanks for raising this proposal. It is a common use case to sink metric
> "data" into downstream Prometheus. The description in the motivation
> section is more or less misleading the discussion. I would suggest you
> rephrase it, e.g. metrics (pre)processing via Flink is ....
>
> The current Flip does not contain too much information about how data will
> be sinked into Prometheus. Would you like to elaborate on it? Some
> architecture diagrams might help us better understand your thoughts. Just
> out of curiosity, since you are focusing on building the Prometheus
> connector, does it make sense to build the Prometheus Source too?
> Short-term stored metrics could be consumed by the Prometheus Source and,
> depending on requirements, might perform some processing like aggregation.
>
> Best regards,
> Jing
>
> On Wed, May 17, 2023 at 6:13 PM Danny Cranmer <dannycran...@apache.org>
> wrote:
>
> > Thanks for the FLIP.
> >
> > I agree that there is a real usecase for metrics sink vs metric reporter.
> > The metric reporters in Flink cover metrics about the Flink job, and a
> sink
> > is used when the metrics are the _data_.
> >
> > +1 on the FLIP ID, can you please fix that?
> >
> > With regard to this FLIP.
> > 1/ The fact we are using the remote write feature is not covered beyond
> the
> > code example. Can we add details on this to make it clear? Additionally
> > would this support _any_ Prometheus server or do we need to enable remote
> > endpoint feature on ther server?
> > 2/ Are there any concerns around Prometheus versioning or is the API
> > backwards compatible? Which versions of Prometheus will we be supporting?
> > 3/ With regard to the "AmazonPrometheusRequestSigner" the example has
> > static creds. Can we integrate with the AWS Util to support all
> credential
> > providers, static and dynamic?
> >
> > Thanks,
> > Danny
> >
> > On Wed, May 17, 2023 at 4:34 PM Teoh, Hong <lian...@amazon.co.uk.invalid
> >
> > wrote:
> >
> > > Thanks Karthi for creating the FLIP!
> > >
> > > Re: Martijn,
> > >
> > > As I understand it, the motivation for the Prometheus Sink is for users
> > > who want to write metrics to a Prometheus remote_write endpoint as an
> > > output of their job graph, so it would be good to treat it as a
> > first-class
> > > citizen and do it as part of Flink’s data flow. This way, we would
> > benefit
> > > from at least once guarantees, scaling, state management.
> > >
> > > For example, a user might want to read in logs, perform some
> aggregations
> > > and publish it into a metrics store for visualisation. This might be a
> > > great use-case for reducing the cardinality of metrics!
> > >
> > > I think you might be referring to the metrics of the Flink job itself
> > > (e.g. CPU / memory metrics). For this use case, I would agree that we
> > > should not use this sink but instead use our metric reporters.
> > >
> > > Regards,
> > > Hong
> > >
> > >
> > >
> > > > On 16 May 2023, at 03:37, Lijie Wang <wangdachui9...@gmail.com>
> wrote:
> > > >
> > > > CAUTION: This email originated from outside of the organization. Do
> not
> > > click links or open attachments unless you can confirm the sender and
> > know
> > > the content is safe.
> > > >
> > > >
> > > >
> > > > Hi Karthi,
> > > >
> > > > I think you are using a wrong FLIP id, the FLIP-229 has already be
> > > used[1].
> > > >
> > > > [1]
> > > >
> > >
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-229%3A+Introduces+Join+Hint+for+Flink+SQL+Batch+Job
> > > >
> > > > Best,
> > > > Lijie
> > > >
> > > > Martijn Visser <martijnvis...@apache.org> 于2023年5月16日周二 04:44写道:
> > > >
> > > >> Hi Karthi,
> > > >>
> > > >> Thanks for the FLIP and opening up the discussion. My main question
> > is:
> > > why
> > > >> should we create a separate connector and not use and/or improve the
> > > >> existing integrations with Prometheus? I would like to understand
> more
> > > so
> > > >> that it can be added to the motivation of the FLIP.
> > > >>
> > > >> Best regards,
> > > >>
> > > >> Martijn
> > > >>
> > > >> On Mon, May 15, 2023 at 6:03 PM Karthi Thyagarajan <
> > > kar...@karthitect.com>
> > > >> wrote:
> > > >>
> > > >>> Hello all,
> > > >>>
> > > >>> We would like to start a discussion thread on FLIP-229: Prometheus
> > Sink
> > > >>> Connector [1] where we propose to provide a sink connector for
> > > Prometheus
> > > >>> [2] based on the Async Sink [3]. Looking forward to comments and
> > > >> feedback.
> > > >>> Thank you.
> > > >>>
> > > >>> [1]
> > > >>>
> > > >>
> > >
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-229%3A+Prometheus+Sink+Connector
> > > >>> [2] https://prometheus.io/
> > > >>> [3]
> > > >>>
> > >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-171%3A+Async+Sink
> > > >>>
> > > >>
> > >
> > >
> >
>

Reply via email to