if the remote filesystem is visible from the other, than a different HDFS
value, e.g hdfs://analytics:8000/historical/ can be used for reads & writes,
even if your defaultFS (the one where you get max performance) is, say
hdfs://processing:8000/
-performance will be slower, in both directions
Thank you guys. I will try JDBC route if I get access and let you know.
On Mon, Dec 5, 2016 at 5:17 PM, Jörn Franke wrote:
> If you do it frequently then you may simply copy the data to the
> processing cluster. Alternatively, you could create an external table in
> the processing cluster to the
If you do it frequently then you may simply copy the data to the processing
cluster. Alternatively, you could create an external table in the processing
cluster to the analytics cluster. However, this has to be supported by
appropriate security configuration and might be less an efficient then c
The only way I think of would be accessing Hive tables through their
respective thrift servers running on different clusters but not sure you
can do it within Spark. Basically two different JDBC connections.
HTH
Dr Mich Talebzadeh
LinkedIn *
https://www.linkedin.com/profile/view?id=AAEWh2