Re: graceful shutdown in external data sources

2016-03-20 Thread Dan Burkert
ndle this issue, including shutdown hooks and a ref counted cache. - Dan On Wed, Mar 16, 2016 at 4:04 PM, Dan Burkert wrote: > Thanks for the replies, responses inline: > > On Wed, Mar 16, 2016 at 3:36 PM, Reynold Xin wrote: > >> There is no way to really know that, because users

Re: graceful shutdown in external data sources

2016-03-19 Thread Dan Burkert
shell shutdown. - Dan On Wed, Mar 16, 2016 at 2:40 PM, Reynold Xin wrote: > Maybe just add a watch dog thread and closed the connection upon some > timeout? > > > On Wednesday, March 16, 2016, Dan Burkert wrote: > >> Hi all, >> >> I'm working on the Spar

Re: graceful shutdown in external data sources

2016-03-19 Thread Dan Burkert
Hi Steve, I referenced the ShutdownHookManager in my original message, but it appears to be an internal-only API. Looks like it uses a Hadoop equivalent internally, though, so I'll look into using that. Good tip about timeouts, thanks. - Dan On Thu, Mar 17, 2016 at 5:02 AM, Steve Lou

Re: graceful shutdown in external data sources

2016-03-19 Thread Dan Burkert
we require the Kudu client to be manually closed so that it can do necessary cleanup tasks. During shutdown the client closes the non-daemon threads, but more importantly, it flushes any outstanding batched writes to the server. On Wed, Mar 16, 2016 at 3:35 PM, Hamel Kothari wrote: > Dan, >

graceful shutdown in external data sources

2016-03-19 Thread Dan Burkert
Manager.scala>, is there something similar available for cleaning up external data sources? - Dan

Re: The difference between driver and master in Spark

2014-04-03 Thread Dan
Can I think driver as client? Driver and master can be located in a single machine or different machines, right? Thanks, Dan -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/The-difference-between-driver-and-master-in-Spark-tp6158p6192.html Sent

The difference between driver and master in Spark

2014-03-31 Thread Dan
Hi, I've been recently reading spark code and confused about driver and master. What's the difference between them? When I run spark in standalone cluster, from the log it seems that the driver has not been launched. Thanks, Dan -- View this message in context: http://ap