We don't use HDFS but GlusterFS which works like your typical local POSIX file
system.
On 11/21/2016 06:49 PM, Jörn Franke wrote:
Once you configured a custom file system in Hadoop it can be used by Spark out
of the box. Depending what you implement in the custom file system you may
think about side effects to any application including spark (memory consumption
etc).
On 21 Nov 2016, at 18:26, Samy Dindane <s...@dindane.com> wrote:
Hi,
I'd like to extend the file:// file system and add some custom logic to the API
that lists files.
I think I need to extend FileSystem or LocalFileSystem from
org.apache.hadoop.fs, but I am not sure how to go about it exactly.
How to write a custom file system and make it usable by Spark?
Thank you,
Samy
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org