ndle this
issue, including shutdown hooks and a ref counted cache.
- Dan
On Wed, Mar 16, 2016 at 4:04 PM, Dan Burkert wrote:
> Thanks for the replies, responses inline:
>
> On Wed, Mar 16, 2016 at 3:36 PM, Reynold Xin wrote:
>
>> There is no way to really know that, because users
shell shutdown.
- Dan
On Wed, Mar 16, 2016 at 2:40 PM, Reynold Xin wrote:
> Maybe just add a watch dog thread and closed the connection upon some
> timeout?
>
>
> On Wednesday, March 16, 2016, Dan Burkert wrote:
>
>> Hi all,
>>
>> I'm working on the Spar
Hi Steve,
I referenced the ShutdownHookManager in my original message, but it appears
to be an internal-only API. Looks like it uses a Hadoop equivalent
internally, though, so I'll look into using that. Good tip about timeouts,
thanks.
- Dan
On Thu, Mar 17, 2016 at 5:02 AM, Steve Lou
we require the Kudu client to be manually closed
so that it can do necessary cleanup tasks. During shutdown the client
closes the non-daemon threads, but more importantly, it flushes any
outstanding batched writes to the server.
On Wed, Mar 16, 2016 at 3:35 PM, Hamel Kothari
wrote:
> Dan,
>
Manager.scala>,
is there something similar available for cleaning up external data sources?
- Dan
Can I think driver as client? Driver and master can be located in a single
machine or different machines, right?
Thanks,
Dan
--
View this message in context:
http://apache-spark-developers-list.1001551.n3.nabble.com/The-difference-between-driver-and-master-in-Spark-tp6158p6192.html
Sent
Hi,
I've been recently reading spark code and confused about driver and
master. What's the difference between them?
When I run spark in standalone cluster, from the log it seems that the
driver has not been launched.
Thanks,
Dan
--
View this message in context:
http://ap