Hi Scott, I think you can do it by catching the exception in the user function and log the current message that the operator is processing before re-throwing (or not) the exception to Flink runtime.
Best, Paul Lam > 在 2018年11月22日,12:59,Scott Sue <scott....@celer-tech.com> 写道: > > Hi all, > > When I'm running my jobs I am consuming data from Kafka to process in my > job. Unfortunately my job receives unexpected data from time to time which > I'm trying to find the root cause of the issue. > > Ideally, I want to be able to have a way to know when the job has failed due > to an exception, to then log to file the last message that it was consuming > at the time to help track down the offending message consumed. How is this > possible within Flink? > > Thinking about this more, it may not be a consumed message that killed the > job, but maybe a transformation within the job itself and it died in a > downstream Operator. In this case, is there a way to log to file the > message that an Operator was processing at the time that caused the > exception? > > > Thanks in advance! > > > > -- > Sent from: > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/