Hi David,

Do you mean when the JobManager starts, the dns has some problem and the
service could
not be resolved? The dns restores to normal, and the JobManager jvm could
not look up the
dns.
I think it may because the jvm dns cache. You could set the ttl and have a
try.
sun.net.inetaddr.ttl
sun.net.inetaddr.negative.ttl


Best,
Yang

David Maddison <maddisonda...@gmail.com> 于2019年11月29日周五 下午6:41写道:

> I have a Flink 1.7 cluster using the "flink:1.7.2" (OpenJDK build
> 1.8.0_222-b10) image on Kubernetes.
>
> As part of a MasterRestoreHook (for checkpointing) the JobManager needs to
> communicate with an external security service.  This all works well until
> there's a DNS lookup failure (due to network issues) at which point the
> JobManager JVM seems unable to ever successfully look up the name again,
> even when it's confirmed DNS service has been restored.  The weird thing is
> that I can use kubectl to exec into the JobManager POD and successfully
> perform a lookup even while the JobManager JVM is still failing to lookup.
>
> Has anybody seen an issue like this before, or have any suggestions?  As
> far as I'm aware Flink doesn't install a SecurityManager and therefore the
> JVM should only cache invalid name requests for 10 seconds.
>
> Restarting the JobManager JVM does successfully recover the Job, but I'd
> like to avoid having to do that if possible.
>
> Caused by: java.net.UnknownHostException: <********>.com: Temporary
> failure in name resolution
> at java.net.Inet4AddressImpl.lookupAllHostAddr(Native Method)
> at java.net.InetAddress$2.lookupAllHostAddr(InetAddress.java:929)
> at java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1324)
> at java.net.InetAddress.getAllByName0(InetAddress.java:1277)
> at java.net.InetAddress.getAllByName(InetAddress.java:1193)
> at java.net.InetAddress.getAllByName(InetAddress.java:1127)
>
> Thanks in advance,
>
> David
>

Reply via email to