[ 
https://issues.apache.org/jira/browse/HIVE-14270?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15395413#comment-15395413
 ] 

Steve Loughran commented on HIVE-14270:
---------------------------------------

Hadoop's FS API is a plugin point: there are more object stores than Hadoop 
itself ships with. I think of google GFS in particular.

I've started on HADOOP-9565 again; this would let you check by saying {{if 
(filesystem instanceof ObjectStore) { ..... } )), though you'd be writing code 
which only worked for up to date hadoop versions.

Anyway, I agree with your point that a config should really go into Hadoop, 
because it should be relevant to more applications. We'd just need to make it 
something with a default ("s3, s3n, s3a, swift, gfs") which things built to run 
against older hadoop versions can use and not get confused on. Not sure about 
S3 though, because AWS s3:// != ASF s3:// 

> Write temporary data to HDFS when doing inserts on tables located on S3
> -----------------------------------------------------------------------
>
>                 Key: HIVE-14270
>                 URL: https://issues.apache.org/jira/browse/HIVE-14270
>             Project: Hive
>          Issue Type: Sub-task
>            Reporter: Sergio Peña
>            Assignee: Sergio Peña
>         Attachments: HIVE-14270.1.patch
>
>
> Currently, when doing INSERT statements on tables located at S3, Hive writes 
> and reads temporary (or intermediate) files to S3 as well. 
> If HDFS is still the default filesystem on Hive, then we can keep such 
> temporary files on HDFS to keep things run faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to