I was able to access Amazon S3, but for some reason, the Endpoint
parameter is ignored, and I'm not able to access to storage from my
provider... :

sc.hadoopConfiguration.set("fs.s3a.endpoint","test")
sc.hadoopConfiguration.set("fs.s3a.awsAccessKeyId","")
sc.hadoopConfiguration.set("fs.s3a.awsSecretAccessKey","")

Any Idea why it doesn't work ?

2015-07-20 18:11 GMT+02:00 Schmirr Wurst <schmirrwu...@gmail.com>:
> Thanks, that is what I was looking for...
>
> Any Idea where I have to store and reference the corresponding
> hadoop-aws-2.6.0.jar ?:
>
> java.io.IOException: No FileSystem for scheme: s3n
>
> 2015-07-20 8:33 GMT+02:00 Akhil Das <ak...@sigmoidanalytics.com>:
>> Not in the uri, but in the hadoop configuration you can specify it.
>>
>> <property>
>>   <name>fs.s3a.endpoint</name>
>>   <description>AWS S3 endpoint to connect to. An up-to-date list is
>>     provided in the AWS Documentation: regions and endpoints. Without this
>>     property, the standard region (s3.amazonaws.com) is assumed.
>>   </description>
>> </property>
>>
>>
>> Thanks
>> Best Regards
>>
>> On Sun, Jul 19, 2015 at 9:13 PM, Schmirr Wurst <schmirrwu...@gmail.com>
>> wrote:
>>>
>>> I want to use pithos, were do I can specify that endpoint, is it
>>> possible in the url ?
>>>
>>> 2015-07-19 17:22 GMT+02:00 Akhil Das <ak...@sigmoidanalytics.com>:
>>> > Could you name the Storage service that you are using? Most of them
>>> > provides
>>> > a S3 like RestAPI endpoint for you to hit.
>>> >
>>> > Thanks
>>> > Best Regards
>>> >
>>> > On Fri, Jul 17, 2015 at 2:06 PM, Schmirr Wurst <schmirrwu...@gmail.com>
>>> > wrote:
>>> >>
>>> >> Hi,
>>> >>
>>> >> I wonder how to use S3 compatible Storage in Spark ?
>>> >> If I'm using s3n:// url schema, the it will point to amazon, is there
>>> >> a way I can specify the host somewhere ?
>>> >>
>>> >> ---------------------------------------------------------------------
>>> >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> >> For additional commands, e-mail: user-h...@spark.apache.org
>>> >>
>>> >
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>
>>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to