2020-02-27 09:18:55 UTC - Prashant Shandilya: @Prashant Shandilya has joined
the channel
----
2020-02-27 09:19:39 UTC - Prashant Shandilya: i m using ansible - python to
setup pulsar cluster .. all works fine .. except my broker keeps on restarting
..
----
2020-02-27 09:19:52 UTC - Prashant Shandilya: anyone faced the similar issue ?
----
2020-02-27 11:45:21 UTC - Yuvaraj Loganathan: Please check and share the logs
----
2020-02-27 16:43:53 UTC - matt_innerspace.io: that makes sense - thanks for the
heads up. wondering if a function, as a consumer, would be slightly more
intelligent, especially if the function is defined for parallelism?
----
2020-02-27 17:31:51 UTC - eilonk: @eilonk has joined the channel
----
2020-02-27 19:26:45 UTC - Addison Higham: :thinking_face: so, just ran into a
case where a developer is looking to do the following with a Pulsar function:
• we have a few hundred topics that map one-to-one to an output topic where we
just do some minor transformation that is fairly consistent
• we don't want to use a function per topic as that is just hard to manage and
expensive
• the problem with doing it in one function though is that it then becomes
really complex to get good perf (gotta use asnyc) but the API model of pulsar
function with it being message at time and not being able to control acking,
that makes it difficult
• two ideas I have:
• 1. Instead of using the context to send, make it so that a pulsar function
returns a `Record` object that the functions machinery can do the work of
sending to multiple output topics
• 2. have an `asyncFunction` that can return a completable future that doesn't
ack the message until the completable future finishes
----
2020-02-27 19:28:33 UTC - Tobias Macey: @Tobias Macey has joined the channel
----
2020-02-27 19:30:11 UTC - Devin G. Bost: @Addison Higham
Can't you listen to the input topics with a regex and then dynamically route to
the output topics based on message metadata (which you could extract)? The list
of output topics could be specified in the function config unless it needs to
be updated more dynamically.
----
2020-02-27 21:50:20 UTC - ikeda: @ikeda has joined the channel
----
2020-02-27 23:16:56 UTC - John Duffie: could use some help with a roadblock -
flink+pulsar+AVRO GenericRecord
We have a special capability where common code operates on GenericRecord
objects. Has no need to know the schema until it arrives just prior to first
message. This works b/c the keys into the record are injecting
programmatically at runtime. Been doing this for years in other environments.
The problem is, the pulsar flink consumer *must* specify the schema before it
connects. That “startup” schema is used for compatibility checks. In our
case, there is no need for a check. Is there a way to get the consumer builder
to ignore the provided schema ?
```PulsarSourceBuilder
.builder(initialSchema)```
Trying to make a rest call to the registry and then use the result when
starting up the consumer in a flink environment is problematic in flink.
----
2020-02-28 00:04:32 UTC - yijie: please try to use
<https://github.com/streamnative/pulsar-flink|https://github.com/streamnative/pulsar-flink>
----
2020-02-28 07:18:04 UTC - Sijie Guo: @John Duffie you can checkout the
connector that yijie points out. that’s the one is contributed to upstream
flink.
----